Friday, August 11, 2023

AI Corner [Humor] : ChatGPT Discusses Senator Warner's Letter to Google re AI & Healthcare

On August 8, 2023, Senator Mark Warner sent a 3 page letter to Google regarding risks and problems that may occur secondary to its release of healthcare-focused AI, like MedPaLM2, to health systems like Mayo.

Find the letter here.

Find an open access article by Katie Adams at MedCity here.

See an essay about the situation below.   I only fed ChatGPT-3.5 the 3-page Senate letter, and asked it to write up a discussion and a headline.

> > > I then asked GPT to discuss the letter in the tone of Lewis Black instead. 


Senator Warns Google CEO About Risks of 

Premature AI Deployment in Healthcare


Senator Mark Warner expresses his concerns to Mr. Sundar Pichai, the CEO of Alphabet Inc. (Google's parent company), about the deployment of Med-PaLM 2, an artificial intelligence (AI) system developed by Google for medical purposes. In the letter dated August 8, 2023, Senator Warner acknowledges the potential of AI to improve patient care but voices apprehensions about the premature deployment of unproven technology. He is worried that such deployment could lead to the erosion of trust in medical professionals and institutions, exacerbation of racial disparities in health outcomes, and increased risks of diagnostic and care-delivery errors.

Senator Warner highlights the competitive rush among technology companies, including Google, to develop and deploy AI models following the increased attention on AI technology, citing the launch of OpenAI's ChatGPT as a catalyst. He also references previous concerns raised about Google's partnerships with hospital systems that involved training diagnostic models on sensitive health data without patients' knowledge or consent.

The senator specifically addresses Google's Med-PaLM 2, which is designed to answer medical questions, summarize documents, and organize health data. Despite these intentions, Google's research has shown that Med-PaLM 2's answers often contain inaccurate or irrelevant information compared to physicians' responses.

Senator Warner poses a series of questions to Mr. Pichai, asking about Google's assessment of potential risks and failures associated with Med-PaLM 2 in a medical context. He inquires about data privacy measures, the documentation provided to hospitals, the frequency of model updates, the training data's privacy implications, patient awareness of AI involvement in their care, and the presence of opt-out options.


  1. Trust and Medical Practice: The letter raises concerns about the trustworthiness of AI-driven medical solutions. Premature deployment of AI in healthcare without proper validation could undermine patients' confidence in medical professionals and institutions.
  2. Racial Disparities: Senator Warner suggests that deploying unproven AI technology could worsen existing racial disparities in health outcomes. If AI systems are not thoroughly tested and validated across diverse populations, they might inadvertently perpetuate biased outcomes.
  3. Diagnostic and Care Errors: Unreliable AI-generated medical information could lead to misdiagnoses or incorrect treatment decisions. This poses serious risks to patient safety and well-being.
  4. Data Privacy: The letter highlights concerns about the use of sensitive health data to train AI models. Unauthorized use of such data could violate patient privacy regulations and erode patient trust in data-sharing practices.
  5. Regulation and Accountability: The letter underscores the need for clear standards, guidelines, and ethical considerations when deploying AI in healthcare. This reflects the growing awareness of the necessity to regulate and hold tech companies accountable for the impact of their AI solutions.
  6. Transparency and Informed Consent: Senator Warner's questions about patient awareness and opt-out options indicate a growing emphasis on the need for transparency in AI use in healthcare and the importance of informed patient consent.

In summary, Senator Mark Warner's letter to Google's CEO raises important concerns about the potential risks and implications of deploying unproven AI technology in the medical domain. It highlights the need for careful evaluation, transparent communication, and responsible deployment of AI systems to ensure patient safety and trust in healthcare practices.


Please write a 150-word discussion of the same Warner letter to Google, but in the snarky humorous tone of a commentator such as John Oliver or Lewis Black, who thinks the letter is paranoid and unnecessary and shows little insight about AI.


Oh, hold onto your stethoscopes, folks, because Senator Mark Warner has penned a letter to Google that reads like a wild conspiracy theory pamphlet. 

Apparently, he's convinced that Google's Med-PaLM 2 AI system is going to unleash a medical apocalypse of epic proportions. Picture this: AI-generated band-aids that randomly burst into flames, or maybe Google's AI becomes a doctor who diagnoses everything as a case of the dreaded "Googloid Fever."

But wait, there's more! Warner's worried about "eroding trust" in medical professionals because of AI. As if trust wasn't already eroding due to WebMD and the entire internet diagnosing us with diseases we can't even pronounce. And don't even get me started on his fear that AI will make medical errors. Because human doctors have NEVER misdiagnosed anything, right?

It's almost adorable how Warner thinks Google is secretly plotting a health takeover. Yes, folks, Google's main goal is to turn us all into hypochondriacal robots. So, Senator Warner, before you sound the alarm on the AI apocalypse, maybe take a deep breath, grab a virtual chill pill, and let the rest of us non-paranoid folks embrace the AI healthcare revolution without envisioning a sci-fi dystopia at every corner.


Warner's letter closes by quoting Singhal et al., Nature (July 12; 620:172, 28pp, "Large language models encode clinical data." More about Singhal.