Industry Watcher’s Digest

Buzzworthy developments of the past few days.

  • A lot of technology leaders have gone out of their way to call for more—or at least smarter—regulation of AI. Add Keith Dreyer, DO, PhD, to the list of those who have long toiled in healthcare and deserve open ears. He’s concerned that current regulatory guardrails are suspect because training data is “changing under the feet of the algorithm.” Last week he went to the nation’s capital to state his case directly to lawmakers and regulators. A radiologist with Mass General Brigham, Dreyer serves as that institution’s chief data science officer. He’s also regarded as a pioneer of AI for healthcare. Dreyer tells Politico that new training data is a clear and present worry because “no one really knows” if updating data makes medical algorithms more accurate, less accurate or an unstable admixture of the two—“and there are no requirements in place to monitor that.” More here.
     
  • Speaking of AI giants with a hand in healthcare, keep an eye on Bill Gates’s brainchild. You know the reference is to Microsoft even though it’s been several years since Gates cut official ties with the Big Tech behemoth that he co-birthed. Anyway, Microsoft is partnering with a full-service revenue cycle management company whose only industry is healthcare. Announcing the development April 23, Ensemble Health Partners in Cincinnati says its proprietary platform, EIQ, will tap Microsoft’s Azure to beef up offerings in automation, machine learning and generative AI “across the entire revenue cycle.” Ensemble announcement here.
     
  • As it happens, revenue cycle management is pretty high on any list of things hospitals want to do with AI. One list it makes is the “Top 12 ways artificial intelligence will impact healthcare” in the estimation of the editors at Health IT Analytics. Read up on AI for healthcare RCM and see if your picks align with theirs here.
     
  • AI startup Hugging Face has put together a way to benchmark the quality of large language AI models for performing specific healthcare tasks. Calling the system Open Medical-LLM, the company says its idea is to standardize evaluations of these models so end-users can know whether or not they should bother. Hugging Face worked on the toolkit with colleagues at the University of Edinburgh’s Natural Language Processing Group. Learn more here.
     
  • Utah’s new AI law goes into effect May 1. Why does this matter to non-Beehive Staters? Because it could set a precedent. And because it has some interesting wrinkles. For example, the law only makes businesses disclose their use of AI when customers ask. However, it doesn’t view healthcare as just another kind of business: Providers outfitted with medical AI have to “prominently disclose” the care component to patients before using it. On the other hand, as long as the patient is informed beforehand, the law doesn’t directly regulate how the AI—generative or otherwise—is to be used. Read a legal analysis from the Chicago-based law firm McDermott Will & Emery here.
     
  • Only 10 AI startups made the cut as finalists. Only one is healthcare-specific. The competition is the 2024 Innovation Showcase of the 21st annual MIT Sloan CIO Symposium. The organization says the 10 have less than $10 million in annual revenues, sell enterprise IT solutions to CIOs or corporate IT departments, and have developed “cutting-edge solutions that combine both value and innovation to the enterprise IT space.” The healthcare finalist is diagnostics streamliner SimulConsult. They’ll be on hand when symposium organizers name the winner/s May 14 in Cambridge, Mass. Announcement here.
     
  • Want a contrarian’s take on the push for more regulation of AI? You got it. Just because people in high tech places want stepped-up governmental oversight doesn’t mean the resulting restraints will be wise or in the public’s best interest. More likely is that OpenAI and other industry players putting out the calls are acting in their own interests: They just want to foil their competitors. “Many regulators will happily support these requests, even when they are being played.” This take is from Eric Goldman, JD, a law professor at Santa Clara University in Silicon Valley. In a paper and presentation unsubtly titled “Generative AI is Doomed,” Goldman makes a dark prediction. It’s worth quoting at length. Here comes:  

“I expect regulators will intervene in every aspect of Generative AI’s ‘editorial’ decision-making, from the mundane to the fundamental, for reasons that range from possibly legitimate to clearly illegitimate. These efforts won’t be curbed by public opposition, Section 230 or the First Amendment. The regulatory frenzy will have a shocking impact that most of us have rarely seen, especially when it comes to content production: a flood of regulation that will dramatically reshape the Generative AI industry—if the industry survives at all.”
 

 

Dave Pearson

Dave P. has worked in journalism, marketing and public relations for more than 30 years, frequently concentrating on hospitals, healthcare technology and Catholic communications. He has also specialized in fundraising communications, ghostwriting for CEOs of local, national and global charities, nonprofits and foundations.