Regulating generative AI needs a new approach. Here is one idea.
Human-like AI agents are coming soon to healthcare, but risks are not fully understood and the regulatory framework is lagging behind the technology curve. Here is a novel approach.
Boost your industry knowhow by staying current with the latest regulatory thinking as part of our Business Acumen series on Let’s Talk Risk!
Imagine a future where you could have a human-like smooth conversation with a virtual healthcare agent about your medications, lab results and even nutrition advice.
They could provide you with nearly as good, if not better, information as a real nurse or a doctor1. Plus you will save both time and money.
Or imagine if a surgeon could receive real-time advice during surgery from a virtual specialist when confronting a life-threatening emergency for the patient.
It could save the patient’s life.
Yes, this is about to become a reality, but the current regulatory framework is not adequate to ensure these technologies are deployed in a safe and secure manner. The old ways of regulating medical technologies is simply not going to work in this brave new world of Generative Artificial Intelligence (GAI)!
Simply put, AI technologies are moving too fast for regulators to keep up. In a recent policy article published in the New England Journal of Medicine (NEJM)2, two thought leaders offer a novel regulatory approach to regulating clinical GAI. In this LTR article, we take a closer look at recent AI developments, current regulatory framework and a novel regulatory approach offered in the NEJM article.
Let us dive in.
Keep reading with a 7-day free trial
Subscribe to Let's Talk Risk! to keep reading this post and get 7 days of free access to the full post archives.