- Towards Interpretable and T...
- Edit Event
- Cancel Event
- Preview Reminder
- Send Reminder
- Other events happening in October 2023
Towards Interpretable and Trustworthy RL for Healthcare
Speaker:
Finale Doshi-Velez
, Gordon McKay Professor in Computer Science at the Harvard Paulson School of Engineering and Applied Sciences at Harvard University
Date: Friday, October 27, 2023
Time: 12:30 PM to 1:30 PM Note: all times are in the Eastern Time Zone
Public: Yes
Location: D463 (Star)
Event Type: Seminar
Room Description:
Host: Marzyeh Ghassemi , IMES, CSAIL, EECS
Contact: Sheila Sharbetian, 617-324-6747, sheilash@csail.mit.edu
Relevant URL:
Speaker URL: None
Speaker Photo:
None
Reminders to:
seminars@csail.mit.edu
Reminder Subject:
TALK: Towards Interpretable and Trustworthy RL for Healthcare
Abstract:
Reinforcement learning has the potential to take into the many factors about a patient and identify a personalized treatment strategy that will lead to better long-term outcomes. In this talk, I will focus on the offline, or batch setting: we are giving a large amount prior clinical data, and from those interactions, our goal is to propose a better treatment policy. This setting is common in healthcare, where both safety and compliance concerns make it difficult to train a reinforcement learning agent online. However, when we cannot actually execute our proposed actions, we have to be extra careful that our reinforcement learning agent does not hallucinate bad actions as good ones.
Toward this goal, I will first discuss how the limitations of batch data can actually be a feature, when it comes to interpretability. I will share an offline RL algorithm that takes advantage of the fact that we can only make inference about alternative treatments when clinicians have tried many alternatives not only to produce policies that have higher confidence statistically but also are compact enough to inspect by human experts. Next, I will touch on questions of reward design and taking advantage of the fact that our batch of data was produced by experts. Can we expose to the clinicians what their behaviors seem to be optimizing? Can we identify situations in which what a clinician claims is their reward does not match their actions? Can we perform offline RL with all the great qualities above in a way that is robust to reward misspecification? That takes into account that clinicians are in general doing their best? Our work in these areas brings us closer to realizing the potential of RL in healthcare.
Research Areas:
AI & Machine Learning
Impact Areas:
Health Care
Created by Sheila Sharbetian at Monday, September 11, 2023 at 12:17 PM.