Academic audience at a University of Pennsylvania artificial intelligence (AI) in health care conference.
The LDI “(Re)Writing the Future of Health Care with Generative AI” conference” opened with Micky Tripathi, PhD, MPP, the Acting Chief Artificial Intelligence Officer of the U.S. Department of Health and Human Services being interviewed by Dan Gorenstein, LDI Adjunct Senior Fellow and Executive Editor and Host of the Tradeoffs podcast. (Photos: Hoag Levins)

As she opened the “(Re)Writing the Future of Health Care with Generative AI” conference, LDI Senior Fellow and Wharton School Health Care Management Professor Marissa King, PhD, pointed out that “there’s already immense penetration of generative artificial intelligence (AI) into healthcare and as we think about how we can harness it in order to improve the quality and efficiency of care, and reduce costs we must be mindful that these incredible possibilities come with a lot of risk. Our program today looks at how we can harness these possibilities while mitigating risk.”

In that same spirit of risk mitigation in the first panel, moderator Dan Gorenstein, LDI Adjunct Senior Fellow and the Executive Producer and Host of the Tradeoffs podcast, interviewed U.S. Department of Health and Human Services Acting Chief Artificial Intelligence Officer, Micky Tripathi, PhD, MPP, and pointed to the risk cited in an article published two weeks earlier in the New York Times. The piece detailed how 150,000 doctors and assistants at more than 150 health systems are using an AI feature in EPIC’s MyChart patient portal to draft their initial responses to patients. The Times pointed out how the “trend troubles some experts who worry that doctors may not be vigilant enough to catch potentially dangerous errors in medically significant messages drafted by A.I.”

Gorenstein further explained that, “A small study in the Lancet found that if left unedited, these messages would pose a risk of severe harm about 7% of the time.” He asked Tripathi to walk the audience through the federal approval process that a product like that MyChart AI feature has to go through before it’s released into the real world of patient care.

“There isn’t really a federal approval process it would have to go through,” said Tripathi. “It doesn’t qualify as a medical device by FDA definitions… And within electronic health record software, there’s no requirement that it go through some kind of approval process.”

Dan Gorenstein, Executive Producer of the Tradeoffs Podcast on stage at the University of Pennsylvania.
To orient the audience to the exact nature of what a large language model (LLM) is, Gorenstein asked ChatGPT to explain that—which it instantly did on screen.
Micky Tripathi, PhD, Acting Chief Artificial Intelligence Officer of the U.S. Department of Health and Human Services.
Tripathi told the audience that he uses LLMs and noted that prior to the event he asked an LLM to tell him what he should be wary of when being interviewed by journalist Dan Gorenstein.
Panel members in a session discussing the ethical issues involved in integrating large language models into health care clinical use.
On the first conference panel were moderator Ravi Parikh, MD, MPP, Associate Professor at the Emory University School of Medicine; Maia Hightower, MD, MPH, MBA, CEO of Equality AI; David Sontag, PhD, CEO of Layer Health and Professor at the Massachusetts Institute of Technology (MIT); and Zachary Lipton, PhD, Chief Technology Officer at Abridge and Associate Professor at Carnegie Mellon University.

Moderating the “Ethical Innovation: Patients, Doctors, and Large Language Models” panel, Ravi Parikh, MD, MPP, of Emory University School of Medicine, noted that the goal of the conference was to “expand a whole spectrum of AI hype and stories of positive implementation of large language models in health care to some more cautionary tales and aspects of regulation and how we can implement in careful ways.”

The panel addressed the complicated ethical challenges involved in the creation and use of LLMs for health care, particularly in relation to Predictive Decision Support Interventions (Predictive DSIs)—the AI-driven systems designed to assist in medical decision making by analyzing data and generating clinical predictions, classifications, or recommendations.

Panelist Maia Hightower, MD, MPH, the CEO and Co-Founder of Equality AI, an organization that provides AI quality assurance and compliance software for health care, emphasized the crucial importance of transparency, despite the enormous complexity of AI systems designed for health care purposes. She noted her strong support for the AI transparency regulations that the Department of Health and Human Services (HHS) Office of the National Coordinator for Health Information Technology (ONC) released in December.

That ONC transparency rule requires that health care AI developers provide detailed information on their algorithms, allowing healthcare providers to assess those systems’ fairness, appropriateness, validity, effectiveness, and safety. Developers must disclose information such as the source of data used for training, performance metrics, validation processes, and how risks associated with LLM outputs are managed.

“We do have this historical problem of clinical decision support tools being unregulated and not providing that level of transparency on where that data is coming from, how it performs across demographic groups, and whether or not health equity was taken into consideration,” Hightower said.

Ravi Parikh moderates a conference about AI in health care at the University of Pennsylvania
The former Director of the University of Pennsylvania Human Algorithm Collaboration Lab, Ravi Parikh is a leading researcher in applying AI to health care, particularly in oncology and advanced illness care.
Maia Hightower, MD, MPH, on stage at a conference on the use of AI in health care at the University of Pennsylvania.
Panelist Maia Hightower, MD, MPH, explained that “Most health systems need help at the very beginning with AI governance and in understanding what AI risk involves, and what it means to be accountable for mitigating that risk, because it is different than application risk or privacy security risk, or [Health Insurance Portability and Accountability Act] HIPAA security risk.”
A panel of academic experts discussing health system's integration of large language model AI systems into clinical care.
Panelists in the “Drafting the LLM Playbook: Key Questions for Health Systems” session were moderator Hamsa Bastani, PhD, Associate Professor and Co-Director of the Wharton Healthcare Analysis Lab; Julia Adler-Milstein, PhD, Professor and Director of the Center for Clinical Informatics and Improvement Research (CLIIR) at the University of California, San Francisco; I. Glenn Cohen, JD, Professor of Law at Harvard Law School; and Sanmi Koyejo, PhD, Assistant Professor and Leader of Stanford Trustworthy AI Research (STAIR) at Stanford University.
Moderating a panel at a University of Pennsylvania conference is Hamsa Bastani, PhD of the Wharton School
“We’re here to ask some of the tough questions about using generative AI and large language models for improving health care delivery,” said moderator and LDI Senior Fellow Hamsa Bastani, PhD. “One of those is who do you think should carry the responsibility for errors these models make, especially when we’re thinking about translating these systems from premier institutions to safety net hospitals?”
“If you’re interested in creative writing applications, the fact that an LLM hallucinates is excellent. It’s artistic,” said Sanmi Koyejo, PhD. “But if you’re interested now in the decision-making setting, this is extremely worrying. And I think part of the gap is that the technology was really not built for the former, and we’re having to come back and patch a bunch of these issues to handle the fact that we need to retrieve correctly and make decisions that we think are accurate.”

Because of the enormity of the planning, implementation, operation, and daily management and quality control of LLM-driven systems, health care systems are going to need a new layer of management headed by a Chief Health AI Officer, said panelist Julia Adler-Milstein, PhD, a Professor of Medicine at the University of California, San Francisco School of Medicine. Her institution has already created and filled that role as part of its AI governance process, supported by both a steering committee and a health AI oversight committee.

“It really is such a scope of work that we need a dedicated person to oversee it,” said Adler-Milstein who is the Director of the Center for Clinical Informatics and Improvement Research (CLIIR). “Our AI steering committee does something really important that hasn’t come up as much in today’s conference, which is talk about what problems are we trying to solve with AI. We almost walked into the room with the assumption that AI is good and doing good—but for what problems and what purposes? What are our health system’s biggest pain points? Which ones might best be solved by AI and how do we do a landscape assessment of all the tools that are out there to figure out if any of them look they’re ready for us to buy?”

“AI is often a mirror, and we don’t always like what we see when we look in the mirror,” said Sanmi Koyejo, PhD, Assistant Professor of Computer Science and Principal Investigator of the Stanford Trustworthy AI Research (STAIR). “But sometimes it’s actually useful when a mirror reflects to us some of the gaps in decision-making that individually seem rational, but collectively suggest bias of various kinds. A Stanford study evaluated a bunch of AI models and asked how they failed in ways related to equity. One quite interesting failure mode involved kidney failure and the EGFR measure that had a race correction and that was later fixed in the real-world literature and science, but not necessarily in the technology. When they evaluated AI models, the researchers found some still used the race correction.”

Asked what AI issue keeps him up at night, I. Glen Cohen, JD, Faculty Director of the Harvard University Petrie-Flom Center for Health Law, Policy, Biotechnology, and Bioethics, said it was knowing that “you can build an LLM system in such a way that it is scalable in terms of the AI, but are the review processes we’re talking about scalable? What happens when you are a safety net hospital and you too want the benefit for your population of these tools you’re hearing great things about? How do we know if the translation to this population is appropriate for the model? How can we do this ongoing review? And how can health systems share information without being sued for defamation when they are saying bad things about a model which produces terrible results for their patients?”

Jenny Ma, JD, of the HHS Department speaks on a University of Pennsylvania conference panel on the regulation of large language model AI systems in health care
On the “Keeping Pace: Regulatory Imperatives for Large Language Models in Health Care” panel were moderator and LDI Senior Fellow Gary Weissman, MD, MSHP, Assistant Professor at the Perelman School of Medicine; Elizabeth Edwards, JD, Senior Attorney at the National Health Law Program; Jennifer Goldsack, MBA, CEO of Digital Medicine Society (DiMe); and Jenny Ma, JD, Principal Deputy Director of the U.S. Department of Health and Human Services.

In the panel on how federal regulations relate to AI, Jenny Ma, JD, the Principal Deputy Director of the Office for Civil Rights at the U.S. Department of Health and Human Services, emphasized how the revised non-discrimination principles of the Affordable Care Act’s Section 1557 will, in July of 2025, be officially applied to health care AI. Section 1557 prohibits discrimination on the basis of race, color, national origin, sex, disability, age, or religion in activities that receive federal financial assistance, including health insurance plans, health care providers, health care facilities, and the AI systems they use.

“These non-discrimination principles are applicable to you if you are a covered entity that has violated 1557, whether you blame it on AI or not,” said Ma. “And we just wanted to affirm that.”

Jennifer Goldsack, MBA, Founder and CEO of the Digital Medicine Society (DiMe), a Boston-based non-profit focused on advancing the safe and equitable use of digital means to redefine and improve health care, emphasized why strong non-discrimination measures were needed in the health care AI field. She pointed to a recent investigative series called “Embedded Bias” in STAT News that detailed how race-based algorithms are already widely used throughout the health care delivery system and why it’s so difficult to change them.

DiMe has been partnering with the New York City Coalition to End Racism in Clinical Algorithms (CERCA) run by the New York City Department of Health and Mental Hygiene. Goldsack said the unusual thing about the project is how CERCA has gone across New York’s five boroughs talking with clinicians who are often able to identify discriminatory aspects of the algorithms responses they are already so familiar with from daily use.

“One of the things we’ve learned is that identifying these things is going to be very little about technology, and an awful lot about how we actually interact with the clinicians and administrators who use the AI. So, when we come to transparency, I want to say the notion that it’s purely a technological issue is not accurate,” said Goldsack. “It’s really interesting to think about how we can create a playbook to create incentives and scale CERCA’s approach and give a reason for every single public health department and every single health system to implement the same thing around the country.”

University of Pennsylvania Assistant Professor Gary Weissman, MD, moderates a panel on the use of artificial intelligence in health care.
Moderator and LDI Senior Fellow Gary Weissman, MD, MSHP, noted that “Earlier today, we’ve heard about hype, use cases, AI models being deployed, and lots of enthusiasm, but we’ve heard less about the actual and potential harms related to the use of AI systems in the health care space. In part, that’s because there is not too much evidence for AI systems in general, and most of the evidence we do have doesn’t touch on issues of safety and harm.”
“The National Health Law Program focuses on Medicaid and low-income health care programs,” said Elizabeth Edwards, JD, Senior Attorney in that organization. “I want to address the comments about how ‘AI just gets deployed and then we fix it later as we identify the harms.’ That is not okay because there are many people using the health system who cannot absorb any level of harm.”

Author

Hoag Levins

Editor, Digital Publications


More LDI News