Big Tech’s race to control generative AI in healthcare raises ethical concerns


Researchers argue that Big Tech should not control generative AI in healthcare.

In an article published in Nature, researchers express concern that big tech companies could dominate the development and use of generative AI in healthcare.

They argue that medical professionals should drive development and deployment to protect people’s privacy and safety – not commercial interests.

Big Tech’s entry into healthcare AI

Tech giants such as Google and Microsoft are making great strides in generative AI for healthcare.



Google recently unveiled MedLM, a set of specialized generative healthcare AI models available to customers in the US through its Vertex AI platform. The models are based on Med-PaLM 2, the second iteration of Google’s large-scale, specialized medical language models that can respond at a specialist level.

Microsoft recently introduced Medprompt, a new prompting strategy that enables GPT-4 to achieve top scores on medical question benchmarks, outperforming specialized models like MedPaLM-2. Earlier this year, Microsoft highlighted the potential of GPT-4 for medical tasks.

Despite these advances, the researchers argue that the rush to adopt proprietary large language models (LLMs) – such as those used by ChatGPT – risks ceding control of medicine to obscure commercial interests. They point to several potential pitfalls.

Healthcare could quickly become dependent on LLMs, which are difficult to evaluate and could be changed without notice or even discontinued if the service is deemed no longer viable. This could undermine patient care, privacy, and safety, the researchers write.

In addition, LLMs often generate hallucinations and convincingly false results. When circumstances change, such as the emergence of a new virus, it is unclear how a model’s knowledge base can be updated without costly retraining.


Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top