
Loading summary
John Miles
Coming up next on Passion Struck, I'm
Dr. Robert Wachter
more optimistic about healthcare AI than I am about the AI and the rest of our lives. Because in health care, I can't find a doctor, a nurse, or a patient who says the system is just perfect and we don't need to screw around with it. It's delivering the high quality, safe, accessible care at an affordable price. Nobody thinks that. And I don't think we can deliver on what patients need unless we have the assistance of a technology like AI. So it really is both the quality of the technology and the desperate need that we have to make things better.
John Miles
Welcome to Passion Struck. I'm your host, John Miles. This is the show where we explore the art of human flourishing and what it truly means to live like it matters. Each week I sit down with change makers, creators, scientists and everyday heroes to decode the human experience and uncover the tools that help us lead with meaning, heal what hurts, and pursue the fullest expression of who we're capable of becoming. Whether you're designing your future, developing as a leader, or seeking deeper alignment in your life, this show is your invitation to grow with purpose and act with intention. Because the secret to a life of deep purpose, connection and impact is choosing to live like you matter. Hey, friends. Welcome Back to episode 742 of Passion Struck. Over the past several episodes in our Life behind the Scripts series, we've been exploring how transformation happens across different layers
Interviewer/Host
of the human experience.
John Miles
With Joan Lunden, we looked at identity, what happens when the roles that once defined you no longer fit. With Harvard Professor Leslie John, we explored authenticity, how revealing more of yourself reshapes connection and belonging. To with biohacking pioneer Dave Asprey, we turn to biology. How your brain, body and nervous system influence the person you're capable of becoming. And in my conversation Last Thursday with Dr. Tara Narula, we focused on resilience, how agency, connection and mindset shape not just how long we live, but how well we live. Today we zoom out even further because there's another force reshaping our lives, not just individually, but systematically. Technology, artificial intelligence is already transforming how we work, communicate and make decisions. And nowhere may its impact be more profound than in healthcare, the domain where technology intersects directly with human vulnerability, trust and survival. My guest today is Dr. Robert Wachter, physician professor, chair of the Department of Medicine at University of California, San Francisco, and one of the world's leading voices on healthcare innovation. We're discussing his new book, A Giant How AI Is Transforming Healthcare and what that Means for our future Dr. Wachter has spent decades studying how medical systems evolve from electronic health records to patient safety to digital medicine, and he offers a uniquely balanced perspective on both the promise and and the risks of AI. In this conversation, we explore how AI could dramatically improve access to care, why burnout and administrative overload have reached a breaking point, the surprising ways AI may restore humanity to medicine the risks of bias, misinformation and deep fakes in high stakes decisions and how patients themselves may become far more empowered participants in their own health. This is not just a conversation about technology. It's a conversation about trust, responsibility, and the future of human care. Before we dive in, a quick ask if this episode resonates with you, share it with someone who cares about the future of healthcare. You can also watch the full conversation on our YouTube channels. And if you haven't yet, leaving a rating or review on Apple Podcasts or Spotify helps more people find these conversations. Now let's dive into my conversation with Dr. Robert Wachter. Thank you for choosing Passion Struck and choosing me to be your host and guide on your journey to creating an intentional life that matters. Now let that journey begin.
Grocery Outlet Announcer
Welcome to Big Savings this week at Grocery Outlet, your Extreme Value headquarters. Right now. Sanderson Farms Boneless skinless chicken breasts are only $1.99 per pound and get one dozen large cage free eggs for only 99 cents. However you cook them up, you're saving big on fresh quality. Stock up on these family favorites today. These deals are only available until March 17 while SUPP last selection varies by store, see weekly ad or in store for grade and size details.
Dr. Robert Wachter
Grocery Outlet Bargain Market.
Interviewer/Host
I am so excited Today to welcome Dr. Robert Wachter to Passion Struck. Welcome Bob. How are you today?
Dr. Robert Wachter
I'm great John, thank you so much for having me.
Interviewer/Host
Well, I've read some interviews you've done and you've joked that you were born in Brooklyn before it was trendy. How did growing up in that second generation Jewish family shape how you see work, success and responsibility?
Dr. Robert Wachter
I was born there and then we moved out to Long Island. Neither of my parents went to college and I guess the funniest story was my dad was really an admirer of doctors. He ran a small business and I remember going he would go to cocktail parties on Friday night and he'd be dressed normally and then I'd look down and he'd have his garage door open on his belt and I said, dad, what's that? He said, there are a lot of doctors at the party. They have beepers I have nothing. So I think it shaped me as having parents who didn't go to college. They very much wanted me to do well, to make a contribution. I think they saw medicine as the highest calling. For me, it was a little bit interesting because I was also a politics junkie and had no idea how I'd be able to shape a career that had my interest in sort of the way systems worked and people interact with each other and also be a doctor and feel incredibly fortunate. Things have worked out pretty well.
Interviewer/Host
Yeah. I understand that you never thought you'd be able to combine political science with medicine, but it's turned out to happen over time. How did that end up occurring?
Dr. Robert Wachter
I met a mentor in medical school at the University of Pennsylvania, a guy named John Eisenberg, who is an MD and an MBA from Wharton. And John was incredibly talented as a physician, as a teacher, as a researcher. He was charismatic. He was larger than life. And what he studied, I thought research in medicine was genes and Bunsen burners and test tubes. And that was. I was not that interested in or nor that good in basic science. But what John studied was the system. How we organize ourselves, how the payment system influences care. Why are so many people uninsured, and how do we fix that? So it was really through John that I said, wow, there is a career here as an academic physician where you study the system of health care. And that was completely in sync with my interest in policy and politics and economics. And I realized, wow, you can be a doctor part time and a researcher part time, and what you can study is this system and how to make it better. And that's given me plenty to do over the last 40 years, because the system really needs a lot of reform.
Interviewer/Host
Yes. So what we're going to be discussing today is your new book, A Giant Leap, How AI is Transforming Healthcare and what that Means for Our Future, which I think is a really important conversation, because I can't think of AI transforming any industry as much as it could potentially transform healthcare.
Dr. Robert Wachter
And.
Interviewer/Host
Or do we need it? But you open the book with. Yeah, you open the book with the Mayo clinic digital twin, Dr. Morris watching his own AI double explain a diagnosis to a patient. When you first saw that, what did it stir in you emotionally?
Dr. Robert Wachter
I found it thrilling. And because I recognize that the system that we currently have has such an extraordinarily hard time providing the access that patients need. And I don't just mean patients who are in rural areas or patients who are in countries that don't have enough primary care. Doctors or cardiologists or oncologists, even patients who are in a city like San Francisco. It's just very hard to get an appointment with a primary care doctor or you have a medical issue and you don't really think you need to see a doctor, but you need to get good information. And so seeing the CEO of the Mayo Clinic showed me this video of a doctor talking to a patient. And the doctor was empathic and quite human and answering questions. And then the big reveal, of course was the real doctors walked in behind this digital twin. And so I found it exciting and really opens up the idea that AI could scale the knowledge of individual doctors or of places like UCSF or where I work, or the Mayo Clinic. After I was done being thrilled by that, I got scared because in some ways the dark side of that is the idea of deep fakes. And that's probably the thing I worry about the most in healthcare. Maybe I worry about the most generally in AI, which is the capacity to create a digital twin of me, which I think I would trust. Talking to you about your health problems also means we have the capacity to create a digital twin of me telling you you should drink lattes to treat your cancer, or you can think of all the kind of absurdities and it looks like me speaking. And so like a lot of technology, it's a double edged sword, but I think the net positives are pretty exciting.
Interviewer/Host
I was listening earlier this morning to a keynote that you gave on this topic. And during that keynote and in the book, you reference Ernest Hemingway's Gradually, then suddenly. And I know for a lot of people electronic health records were supposed to be a healthcare sudden moment, but they weren't. Why does AI feel like it might just be?
Dr. Robert Wachter
I think there are a couple of reasons. Hemingway's line about how does a man go bankrupt gradually and suddenly feels perfect. Because we certainly have the gradually part down pat. If you think about it, healthcare, maybe, with the exception of education, healthcare is the last industry to be transformed by technology. Just think about the way you access entertainment on Netflix or Amazon prime or whatever. In San Francisco, I take a Waymo. I take a car with no driver about once a week, I'll sit in the backseat, sing Springsteen at the top of my lungs, and sometimes take a nap and completely trust this thing will get me to a place that's safe. So whereas in healthcare, we still are the biggest users of fax machines, we finally ditched the beepers, but we're the biggest users of fax machines in the universe. So it has been gradually. Why suddenly? I think two things. One is this AI is really quite remarkable, and it has a talent that no priority AI had, which is that it can read unstructured data or hear unstructured speech and understand it, or quote, understand it's not human, but take it in, process it, and then provide outputs that are really remarkably human like. And in medicine, that's quite important. We had AI up until before three years ago. It could analyze computable data. It could take your hemoglobin number or your creatinine number or all that and make sense of that, but it couldn't analyze my note about your visit. And now this new AI can. So the capacity of it to do really amazing things is there. But I think the. Probably the more important reason is that the healthcare system desperately needs it. And that's why in some ways, I'm more optimistic about healthcare AI than I am about the AI and the rest of our lives. Because in healthcare, I can't find maybe, you know, some folks, John, but I can't find a doctor, a nurse, or a patient who says the system is just perfect. We don't need to screw around with it. It's delivering the high quality, safe, accessible care at an affordable price. Nobody thinks that. And I don't think we can deliver on what patients need unless we have the assistance of a technology like AI. So it really is both the quality of the technology and the desperate need that we have to make things better. If you look at healthcare over the last 20 years, the way we have dealt with unmet needs is we've just hired more humans. See, not enough doctors. All right, we need. We'll hire nurse practitioners or physician's assistants. Patients with cancer have incredibly challenging pathways where their care needs to be coordinated. Okay, we'll hire a boatload of care coordinators. Dealing with insurance companies is complicated. We'll hire 1,000 people in my hospital to deal with the insurance companies and write authorization requests and all that. It's. We can't do that anymore. We can't afford it, and we can't even find the humans anymore. So it's really a combination of, really a breathtaking technology, but also an industry that absolutely needs the help.
Interviewer/Host
I don't think I would have thought a decade ago I would ever say this, but I think the health care system could actually look at what I've experienced at the VA as an upgrade. And where I'm going with this is, yeah, funny thing to say, but we have this system called my healthevet, and in it, I, as the patient, can look at my complete medical record. I can look at all my lab results. I can see images that are taken even, which is great when you have a wife who's a pcp, because you can see the tests that were done, from MRIs to blood tests, et cetera. But what I really like about it is when I go to community care outside of the system, you can also upload all of those images and files, and so your primary care provider within the VA has access to all of that, so they can clearly see the complete patient arc. But that is such an anomaly compared to what most people experience where they can't see their record go from one department to another, and they're certainly not seeing their test results.
Dr. Robert Wachter
I guess let me just say, John, that I think that has improved a fair amount. Most people in the United States now get care outside of the va, get care in systems that have an electronic health record made by a Wisconsin company named Epic. And if you have Epic, you have MyChart, which is a version of my Healthevet. It is a patient portal. And the problem is the VA is one uniform system across the country. And so if you go to an ER a thousand miles away, if it's a VA er, the information will be there. In epic, it probably won't be, or it's a little harder to access. But in both cases now, I think the patient portal is a real step forward and has democratized care. The problem, as we think about AI, is we democratized care, but didn't give people any of the tools that they need really to use it. So patients now are seeing, well, my magnesium is low and my TSH is high, and my EKG is abnormal, with absolutely no context or no information to help them interpret those things. And AI will help them do that, as your wife will tell you. I'm sure the other thing that the patient portal has done is in addition to having people have all this access to this information, but really no tools to deal with it, or even basic tools to schedule an appointment the way I can schedule a haircut or a dinner reservation. It's given them just enough to confuse them or give them the illusion of democratization. But none of the tools and EPIC and the other EHR vendors have done something that I think everybody thought was helpful, which is put a little button on the bottom of the screen that says, send a message to your doctor. And so patients, being rational human beings, do that. They click that button. And so someone like your wife not only has her 8 or 9 or 10 hours a day of seeing patients. She comes home at night and after dinner or putting the kids to sleep, now has two hours on average of work, not only completing her charts, but ANSWERING on average 50 of these questions a night. It's not doable. And I think AI has the promise of both interpreting all of this data for patients and I think answering a fair number of the questions. So it's really on top of this democratization. We just need tools like this to make the whole thing work. Yeah.
Interviewer/Host
So, Bob, I understand you were once optimistic that electronic health records would transform medicine. And then what they've ended up doing, as you're describing, is they become one of the biggest drivers of physician burnout. Because it's every.
Dr. Robert Wachter
It's.
Interviewer/Host
It's like EPIC brought every single attorney, HR specialist, and quality control person into the room and asked them, what questions should we ask a doctor to answer?
Dr. Robert Wachter
What checkboxes should we put on the screen to make sure the doctor is spending all of his or her entire time looking down at the computer, checking rather than looking you in the eye?
Interviewer/Host
Yeah. And it leads to this question. You write that healthcare's digital failures had shockingly little to do with the technology itself, but more about what I'm describing. What were the real failures and how do we prevent them from happening with AI?
Dr. Robert Wachter
It's a terrific question. The thing that the epiphany I had when I wrote a book 10 years ago called the Digital Doctor about the electronic health record, and this phase was that this was not so unusual the history of what are called general purpose technologies. These are new technologies that transform everything about the nature of work. And so examples of this were electricity or the steam engine or the automobile or the computer or the Internet. And now AI. The tradition is something called, or the paradigm is what's called the productivity paradox of it, of information technology. The paradox is it looks great in the demos, it looks great on the PowerPoint slides, and you bring it in and it does not achieve what you had hoped for and been promised. The reason is that two reasons. One is the technology is not very good in version one and has to get better as the companies iterate on it. But the most important thing is that industries and human managers typically bring the technology in and don't do the hard work of transforming the way work is done. And that just means we're human. We don't have the imagination that we need to understand what our work with technology is going to look like until the technology is actually in our world. This is an old problem in technology. Henry Ford was reputed to have said, if I asked people what they wanted, they would have said, faster horses. They had no ability to think about what cars would do for them or their lives. And until there were cars. And I think the same thing is true here. The electronic health record, we brought it in, we turned it on, we didn't really think for a moment about, okay, what's going to happen when patients have a portal with all this data but they don't understand it? And we have a little button that says, send a message to your doctor. Of course they're going to click that button. We had no plans for what does that do for the workforce, the workflow, how does payment work, all of those sort of things. In some ways it was a natural set of mistakes and our optimism was in some ways misplaced. I think from AI a few things have happened. One is we've learned the importance of anticipating, to the degree that humans can, how to change the nature of the work. We've built up governance structures in healthcare organizations and in doctor's office to figure out how to vet the tools. We're not as gullible as we were to the pitches. We want to see evidence about whether it works. We want to be sure that clinicians were involved in the development of these tools. I think most of our organizations have built the kind of governance that we needed for other digital tools, but now is more ready to deal with AI. The AI is just easier to deal with. You don't need to know a whole lot of technology in order to use it effectively. If you think about the way you use GPT or Gemini, you don't need to be a computer nerd to use it effectively. So I think we're, and as I say, I think the needs of the healthcare system are so vast that I think we're likely to embrace it. I guess the final thing is there's some politics here and the politics are that if the incumbents, particularly the most powerful ones, fear for their jobs, and of course everybody's a little bit worried about what's going to happen to my job with AI. But if the incumbents fear for their jobs and they won't welcome AI the way it would be great for them to. But in healthcare, I think for the foreseeable future, the jobs of doctors and nurses are going to be safe. There's a lot of administrative people and back office people that may, I think, be challenged in terms of their jobs, but I think we're being smarter about it this time. I think most doctors and nurses are going to welcome it into their world because they know how hard their job is and how they can't deliver to the patients what they want to deliver without it. I'd say maybe one final thing is in a technology this complex and in some ways this fraught, you want to start with low hanging fruit. You don't want to start on the hardest problems. And in medicine that means don't start with the AI making a diagnosis for you or recommending a complex treatment for you. Start with a digital scribe where the AI will record my conversation with you as a patient and create my documentation so I can look you in the eye when you're coming in my office. Start with summarizing the patient's old chart. One out of five patients has a medical record longer than Moby dick, longer than 600 pages. I can't possibly review that in detail in three minutes before I see you. So that is something AI can do well. Start with the things that are quick wins to get the kind of buy in that you need. I think those are all lessons that we didn't understand with the electronic health record, and I think we're doing a better job this time with AI.
John Miles
I want to pause for a moment. One of the core ideas in this series is that rewriting your life isn't only about personal insight. It's also about navigating the systems you live within. Your health system, your work environment, your technology landscape, your community. These forces shape what's possible long before we consciously notice them. On theignitedlife.net, i'm sharing companion reflections and articles for each episode in this series designed to help you think more deeply about your own life. Because awareness creates insight, but intentional action creates agency. If you want to explore the reflections for this episode, you can visit my substack@theignitedlife.net thank you for supporting those who support the show. Now a quick break for our sponsors. You are listening to Passion Struck, airing on the Passion Struck Network. Now let's return to the discussion with Dr. Robert Wachter.
Interviewer/Host
I wanted to just make a couple comments. I spent most of my career before I got into what I'm doing now in technology. So I was the CIO of a Fortune 50 company. And so I understand what you're saying about how most people equate technology. Change is going to radically revolutionize whatever it is they're doing. And what everyone fails to realize is it's the cultural change, the people change. That is the biggest burden that any of These implementations have to overcome. And after I left that Fortune 50 company, I went to a company that was in the healthcare space and I had been a long term employee at Lowe's and we had built this system called total closed loop where we were trying to give a Lowe's customer a total closed loop experience regardless of whether they came in through the store, through the website, through the call center. And we were trying to implement this in healthcare. And as we got into it, we realized how problematic it was because the cornerstone of any of these transformations is trying to get your arms around data or what we call metadata in it. And that is so difficult to do in the healthcare ecosystem because the sources of data are so distributed that there's not a single one. Hopefully that's changing. But that was one of the most difficult things to get your arms around.
Dr. Robert Wachter
Getting a little bit better, but not still hard. And also if you think about it, in a corporate world, your interactions are mostly with an individual consumer. And if you convince them that you have a product that's terrific and the right price, they'll buy from you. The number of middlemen in healthcare that you have to deal with in this as you try to close the loop, it also needs to include the insurance company, your employer, the government. It's just the level of complexity is so much higher than people are used to if they've been in more consumer facing industries.
Interviewer/Host
So since he just left his illustrious career, I've got to bring Warren Buffett into this conversation because you quote him saying that healthcare is the tapeworm of the U.S. economy. So it begs the question, does AI meaningfully change that equation or does it just digitize the parasite?
Dr. Robert Wachter
That's a great question. I love that quote because it's absolutely true. If you think about it. I'm old enough to remember when healthcare was 12% of the GDP and people said this is unsustainable. It's sapping up so much of our resources. Well, they're obviously wrong because it's about 20% today and at 20% of GDP it just takes a larger and larger portion of business profits or businesses ability to invest in to do R and D in government funding and therefore government's ability to either keep taxes low or house the homeless or provide decent public education. All of those things are true. Will AI save us money? I think is in some ways one of the unanswered questions. I'd say if you looked at the last 20 or 30 years of health care, you would be skeptical that it will. In every other industry when we innovate with technologies, we usually lower costs and lower prices. In health care, when we innovate with technology, we. We somehow figure out a way of not lowering prices and, or lowering overall costs. Now, part of that is the demand curve is elastic. And what I mean by that is as Lasik surgery for your eyes or replacing your hip becomes more affordable, then more people want it and more people get it. And therefore, even if the net cost of that procedure goes down, the overall cost to society goes up. I think the other. There's another difference, which is health insurance. Nobody really cares. Nobody talks about the percent of our GDP that we spend on computers or on Ubers or on restaurants. People make purchasing decisions. They make their own choices about those things. Whereas in healthcare, because in so many cases you're not paying out of pocket, it's coming out of some public pool, whether you're businesses or government. The ways it could decrease costs. So I'm a little hopeful the ways it could decrease costs. Our two major costs are labor. And so if we need fewer humans around to do the things that we need to do, or if you're able to get care yourself through an app, rather than needing to see a doctor, that could lower cost. The second is that it's been said that the most expensive piece of technology in a hospital is the doctor's pen. Now it's not the doctor's pen anymore, it's the doctor's fingertips. Because what doctors make is about 8% of the healthcare dollar, but the decisions that doctors make is 80% of the healthcare dollar. So a way it could lower costs is if the AI gives me decision support and says, oh, it looks like you're about to order drug A for this patient's cancer or cholesterol or diabetes. Turns out drug B works just as well and is half the price. Don't you want to do drug B, doctor? Or maybe you have to do drug B, doctor. That could lower costs. On the other hand, I could also see a scenario where it says drug B is a little bit better than drug A and costs an extra $50,000 a year, and we recommend you use the more expensive drug. So a lot of that is going to be the devil's. It's going to be in the details of how do those recommendations get framed, whose values are underlying the recommendations. Because sometimes that's a tough decision about whether you recommend the more expensive therapy for a small marginal gain. And the AI is going to have embedded in it some human's logic or some business's logic. About what recommendation it gives you.
Interviewer/Host
Thank you for explaining that. And I want to get more into the power and the pitfalls. But because this conversation has been a little bit deep, I thought I'd bring in a little bit of humor. So I understand that you once played Elton John in front of 2,000 people and were a college mascot. If I have my facts correct, those are both. How important has humor been in surviving in a high stakes professional life like you have?
Dr. Robert Wachter
I think it's absolutely critical. I completely believe that it's vital a not to take ourselves too seriously to try to find joy in work, even the work I do. I just finished 10 days on medical service at UCSF taking care of very sick patients. Several of my patients died incredibly tragically. And yet you need to figure out, I don't think you can survive the work or be very effective if you're not a whole human being. And a whole human being, I think, includes being able to laugh at the things in life that are funny or ironic. And there are many of them for me. Also, I'm chair of a very large department of medicine at a spectacular place. I have a thousand doctors that report to me. I never saw myself as a leader because I always thought I was too goofy. And you've given two examples of my goofiness. And now I've realized, and I had a little imposter syndrome the day, my first day as chair. And then I realized actually the job of leadership today, a lot of it is about authenticity and accessibility and empathy and understanding where people are at. And I think humor is absolutely critical in being the kind of leader that people find accessible. And I had a view of leadership of these people chiseled into Mount Rushmore. And I just don't think that's effective leadership anymore.
Interviewer/Host
So I'm going to just detour one more time from our topic. A lot of what I discuss is around the topic of human flourishing. And I'm. I've just been in over a year writing a book on mattering and how to restore mattering into society because I think it's been slowly being stripped from us. And one of the biggest areas that mattering seems to be stripped is in the workforce. I mean, you can see this in employee engagement, social media surveys. So being in your position in, in charge of so many doctors, what do you see as one of the clear gaps in the workforce that might be causing people to feel this loss of significance?
Dr. Robert Wachter
Yeah, and it's shocking to me when I feel that in my world, because I can't Imagine a job or a workplace where things are more set up for flourishing and for people finding meaning in the work. Academic medical centers are remarkable animals because the law schools don't run a law firm, the business schools don't run businesses. We take care of patients, we teach, we do research. It's fundamentally interesting and important. Where it's being sapped from our workforce, and this is an area where I'm hopeful that AI will help is people spend so much of their time doing kind of bureaucratic paperwork that adds no value to the patients. And we all realize that this is not. It's not useful time. And over time, it just saps people's energy. It takes them away from the meaning that brought them into the field in the first place. I'm hopeful here that AI will make a big difference. I think that one of the things it does really well is some of the bureaucratic hassles. It's been estimated that about 30% of healthcare dollars are spent on kind of administrative paper pushing. That really doesn't add. And I'm not arguing that there should be no regulations there. There needs to be. What we do is very high stakes. We gotta get it right. But it is just layered and layered. And the amount of paperwork that we do, it can overwhelm you in your day. So that's the part I'm hopeful about, that this will take some of that away. If you can. If for your wife as a primary care doctor, if instead of spending two or three hours a day answering emails or writing to insurance companies or keeping her head down checking boxes, she can take that time and repurpose it by looking at you in the eye and really being present for her conversations. I'm guessing her work life gets substantially more satisfying.
Interviewer/Host
That would be fantastic.
Dr. Robert Wachter
Yeah.
Interviewer/Host
For all her and so many other doctors.
Dr. Robert Wachter
Yeah.
Interviewer/Host
So I want to talk about some of the potential pitfalls of AI. So I wanted to bring in a scenario. I used to live in Austin, Texas, before I moved to Florida, where I live now. And good friend of mine is a radiologist. And I remember he would. He was a partner in a pretty big practice, but like once a month or every six weeks, he would be on call over the weekend and he would be doing a lot of reading of scans that would come in. And I was blown away because I asked him, so how many scans do you do in a day? And I was thinking it was a few dozen. And he's, no, it's like 1500 to 2, 2000. And I said, how in the world do you do that. And he said, well, we have a team in India who are doctors who are reviewing these before they come to me. And so they're highlighting on a chart what I should look at. But to me, it's a scenario that you could potentially replace that doctor in India with AI and have AI do these screenings that a human was once doing. But in both cases, he was telling me that even if it was a physician in India looking at a chart, he would still look at that scan himself to make sure they didn't miss anything. And that's my big question. I can see this becoming a category shift, but I also see it being really complex because AI could miss a diagnosis of a potential tumor in one of these scans.
Dr. Robert Wachter
So I got bad news for you. So could he.
Interviewer/Host
So could he.
Dr. Robert Wachter
As I quote Biden in the book, it's often said, don't compare me to the Almighty, compare me to the alternative. And I think when we look at these AI versus humans versus AI plus human scenarios, we have to start with the recognition that humans are not perfect. And we get it wrong a fair amount. And so if our standard is the AI has to be perfect, then I think we're going to miss out on a lot of opportunity. I think it doesn't have to be perfect to be better, it just has to be better. And I guess my assumption is again, I'm influenced a little bit by the fact that I take a driverless car once a week. A thing incredibly high stakes. If you've ever been to San Francisco, making a left turn across Divisidero Street, I can't imagine something more complex than that. And so far, in many millions of miles driven, there's not been a fatal accident involving a Waymo. And I think that it's just a matter of time and it's already happening that AI will read, do, at least do that first read of that scan and replace that doctor in India and probably do it at a lower cost than the doctor in India. It raises a whole different set of questions, which is the propping up the middle class in many other poor countries has been through efforts like these that may change. But yeah, I do think that for the foreseeable future we're going to need what's often call the human in the loop, in this case the doctor or the radiologist in the loop, to do the final reading, the final oversight, because the AI is not going to be perfect and we're not going to completely trust it. So to me, in some ways, a fundamental challenge for this stage is the AI is going to be good enough to be useful, but not perfect. And therefore the doctor is going to need to look over its shoulders. I think that's probably going to be a better system, a more efficient system, a system that can deliver high quality, accurate readings at a lower cost than what we have now. There are some flaws in it, and I'm guessing if your friend, the radiologist, gets 100 scans in a row where the AI has been perfect, the idea that he's looking really carefully at scan 101, just as opposed to signing off on it, he's a human being. He probably is not. There's a second problem, which is that we're now beginning to see studies where the AI is so good that when the human then weighs in, they actually make things worse. They actually degrade the performance of the AI. And so we've got to be sensitive and attuned to that. And the third problem is what's known as deskilling, which is even if your friend stays alert for scan 101, over time, he may be less good at it than he was in the beginning. And there was just a study that came out last year of gastroenterologists looking at doing colonoscopies. And for three months they had access to AI enabled colonoscopy. It puts a little box around the lesions inside your colon and they liked it and it improved their performance. Then they turned it off and their performance for the subsequent three to six months was lower than it had been before they used in the first place. And these were not newbies. These were people who'd been doing this procedure for an average of 10 years. I think it's going to be a huge number of interesting, complex questions of how do we get this dyad of AI plus humans? Right? But I think that your friend, the radiologist is going to eventually feel like this is actually helping him do his job, able to do more scans and do them better. The issue of course, will be at some point, when does he get ditched? And the average US radiologist makes about half a million dollars a year if the AI can read the scans as well as he can. But it costs $50,000 a year to bring into a system, he probably is. He's probably going to be out of a job. I don't think that's happening anytime soon, but hard to believe it doesn't happen within 10 or 20 years.
Interviewer/Host
Eye opening. Now, the next area I wanted to go to is when we think about some of the major social media networks, bias and misinformation have become huge issues. And when I think about medicine, and it's something you touch on in the book, medicine is uniquely vulnerable to AI overconfidence and to biases and that misinformation. So what makes those things more dangerous in healthcare than in almost any other field?
Dr. Robert Wachter
Well, the stakes are incredibly high. The emotions are high in healthcare. Our costs are unbelievably high. There are very few other fields where the frontline practitioner, let's say you're a lawyer or an accountant or a businessman, I could easily, through the stroke of a pen or a keystroke, prescribe a medicine or order a procedure that'll cost $100,000. So the stakes are high, both in life and death and in money. I think the bias issue is we're particularly vulnerable in medicine in the corporate world. You're selling to people that are paying for whatever you're selling out of their pocket, by and large. And if you're a company, you probably have no aspiration to sell to all comers. There's a population who is eligible to buy the thing that you're selling and those are the people that you're trying to cater to. You walk into my er, you will see millionaires and you will see homeless people. You will see people who are English is their, is their fourth language. And the AI tools that we might build for a healthy 30 year old have to also work for a 80 year old person who speaks only Mandarin and lives in a third floor Chinatown walk up. Those are, that's a pretty hard thing to do. The biases are also very real. There have been many studies in health care that show that if you take the same patient who comes in the ER with a fracture, if the patient is white, they get more pain medicine than the patient is black. Now that means humans are biased. But if AI takes what it sees in the medical record and says that is the right way to deliver care, it basically encodes the biases in its software and therefore it scales it. I think the real issue of a lot of the things in medicine are we can hurt somebody one patient at a time. If the AI gets it wrong. Let's say a psychiatrist who's really bad can maybe hurt 100 patients or a thousand patients. A chatbot doing mental health that's really bad could hurt a million patients. So the scaling I think is really a problematic issue. But again, compare me against the not to the almighty. These are, many of these are human problems. Now that we have the capacity to scale it, we've got to think about them in a very different way. Way.
Interviewer/Host
I've had Professor Ethan Mollock from your alma mater, Pennsylvania on this show, and you quote him in the book, the AI you're using today is the worst AI you'll ever use. Do you find that helpful or terrifying in medicine?
Dr. Robert Wachter
In medicine, more hopeful than terrifying. First of all, I think Ethan's one of my heroes. I think he's. He is really more one of the most thoughtful people in what AI is going to do to every industry. Yeah, again, I think the stakes of medicine are so high, the work is so hard, that having it become better and better to me has to be a net positive. Now, where is it potentially terrifying? I like my job. If it becomes so good that it replaces a doctor, that may be good for patients. It may be that patients can get more accessible, more convenient medical care at a lower cost because I'm probably more costly than the bot is. So it could be harmful for different sectors of society if it begins taking over more jobs. But if you look at it from the perspective, what is the healthcare system here to do, which is not really to provide full employment, it's to take care of people and do it at a cost that society can afford that I'm hopeful about. Where I get scared is things like deep fakes. The idea that anybody can be made to look like they're saying anything, I think creates a level of potential distrust, which I think is going to play out in everything, in politics and in every other walk of life. And it'll have its own particular flavor in medicine. And I think the American public is with me. The Gallup survey last year asked people how they feel about AI, and they were very negative about what its impact is going to be on politics, unemployment. The only area they were positive about was in health care. And I think they're right. So I'm looking at this through the lens of healthcare. Whereas the AI gets better. I think it's get better, it's better and better for healthcare. I think for the rest of society, I'm actually quite fearful of it. It makes it easier to design a bioweapon, for example.
Interviewer/Host
So as I mentioned earlier, one of the things I really concentrate on is human flourishing. And I have a lot of friends who are in the medical profession, and I've probably not seen a group of friends that I have that are more burned out than doctors. And so many of them feel indifferent. And it's almost like they're losing their humanity because of the way that the system is working. Right now, is it fair to say that AI could be the first way that more of those doctors could get their humanity back?
Dr. Robert Wachter
I think so. I think you have to disentangle a few different issues that are going into their burnout. The issue of the kind of administrative paper pushing, which is a real issue, I think is very important. When we rolled out these AI scribes to create the documentation at ucsf, the feedback from the physicians was ecstatic. People said I was thinking of retiring and I'm not anymore. So I think that's low hanging fruit. I think that the paperwork, the dealing with the insurance companies and back and forth, I think those things are relatively straightforward and I think will help quite a bit. The other parts of this though, I think there's some more fundamental problems that are going into clinician burnout that I'm not sure AI is the answer to. It used to be that physicians were uniformly trusted and respected. Think back to the days of Marcus Welby and Dr. Kildare. There's a lot of skepticism of expertise in general. And I think this played out during COVID and I think physicians feel that physicians went into this field because they want to help people. And having patients come in being quite skeptical that what you're talking about or pushing back on evidence based recommendations is frustrating and you get tired of having those arguments. I don't know that AI helps that. And in fact, with the world of deep fakes combined with social media, I could actually see that getting worse. So there are parts of this that I think make the job of being a doctor or a nurse better. A lot of it is the paperwork part, but the issue of trust and there's also some built in conflict in the medical system that I don't know AI is going to completely fix. And that is because of the third party payers. Again, when you're going to buy something from the store, it's you and the store and you make a decision whether it's worth it. Here. As we get better and better therapies for obesity or for Alzheimer's or for cancer. But these things cost tens and sometimes hundreds of thousands of dollars a year. There is going to be an inevitable push pull between patients who want a thing and insurance companies or payers who don't want to give you that thing. And physicians find themselves often right in the middle of that. And it's a really uncomfortable place to be. I don't know that AI fixes that.
Interviewer/Host
I want to talk about the other side of this. So there's the physician side and then there's the. The patient side. So one of the big things I found in my own life and I advocate to other people is becoming the CEO of your own health journey because no one knows your body better than you, no one can see all the information better than you. And I think AI can help people become more empowered patients. But it does change in some ways the power dynamic in care. Do you think it does it for better or for worse?
Dr. Robert Wachter
I think for better. I'm a big believer in the democratization of care. I think when a patient comes into the office, assuming I have time, which is its own issue, and they've done some preliminary research and the research they can do today, putting their information in GPT or Gemini is much better than they could have done three years ago. Putting it into Google, I think that's good. I think they're more informed, they're more engaged. I think that's net good. I think where we're going to get in trouble is that the tools right now aren't great for patients. And what I mean by that. The Washington Post did a piece a couple of months ago where the digital journalist Jeff Fowler asked me to review 12 interactions between patients and GPT on healthcare. So they had questions about their health. A lot of the things that GPT said were terrific and one of them actually was far better than what I would have said. It understood some new nuance in the literature that I wasn't aware of. But some of them were awful and one of them was shockingly bad. And I actually potentially, I thought even fatal. When I put stuff into GPT or Gemini and get an answer back. I'm an expert, I can look, first of all, I know what to put in. And second of all, I see the answer and I say that's actually quite smart. And that other thing it's telling me is ridiculous and I'm not going to listen to. Patients have no ability to do that. I watch my daughter go through med school. That process is basically going from layperson to expert. And so the tools that patients have to use AI to help them essentially do self care, I think are somewhat limited. And I think they'll get better. I think I can envision a tool where it's not you just put all your history into GPT, but it's you say GPT. I woke up this morning and I'm dizzy or I'm short of breath. And then the tool says to you, well tell me. Here are a bunch of questions I need to have the answers to in order to understand it. Do you have a fever? Are you having chest pain? And you have a back and forth discussion as you would with a doctor. I think that would make patient facing AI much more useful. Using generic tools like Gemini and GPT I think are pretty imperfect for that function right now.
Interviewer/Host
One of the comparisons that you made in the book is that this moment in time is similar to photography replacing portrait painters. Well, when technical mastery fades, what remains essential to being a doctor?
Dr. Robert Wachter
I guess we'll see. Because I think we used to think empathy and yet these tools are capable of faking empathy pretty darn well. Which for those of us who are hanging on to your question like I went to 10 years of school, what is it that's unique about me? I think it's a tricky question. I would not want a bot to tell me that I had a really dreaded disease. I, I think that's gonna be really important now. Is that forever? I don't know. But for the foreseeable future, if I have a new diagnosis of kidney failure or diabetes or cancer or AIDS or you name it, I would wanna know that from a human. And I want that interaction. I wouldn't want the AI to make a high stakes decision for me, the decision to start chemotherapy or take me to the OR for surgery. I want a human involved in that. For many patients, simple care of and use simple advice a little bit, but I think correctly, the care of your blood pressure, your care of your cholesterol, the care of your diabetes, once we know you have the diagnosis, I think actually could be done by AI and done well, safely and more conveniently and probably less expensively. But a lot of patients have five things wrong with them and I wouldn't trust AI to get it right because when you have five things wrong with you and you have a new symptom because. Cause there are a lot of interactions between those things that are hard to parse. And I think humans are better. I think a lot of patients need a lot of fairly complex team care and I think humans are better at coordinating that and interacting with a social worker or pharmacist to try to get you the care you need. So I think there's a fair amount left for doctors to do and I think there will be for the foreseeable future. Which is partly why I don't think there are going to be unemployed doctors in my lifetime. But I do think doctors are going to be able to augment what they are able to do. Probably see more patients, manage more patients in a way that still is more satisfying. And better for the patients.
Interviewer/Host
I have two last more fun and philosophical questions for you. You've spoken beautifully about how your parents have given you freedom to define success. How has that shaped how you have parented and how you mentor younger physicians today?
Dr. Robert Wachter
Parented? I think I've been really quite open to my kids doing whatever it is, it felt right for them. And you know that. And one of my kids is mildly autistic. And what is success for him is a very different version of success than for my other son or daughter. One of my sons works in Major League Baseball, does baseball analytics. Another is a doctor. My younger son works in an escape room downtown in San Francisco. And success to him is his ability to greet you when you come in and be the voice of God when you're unable to get out in this puzzle. He's never going to be the president of anything or the CEO of anything, but he's wildly successful because he's really achieving his full potential and doing it with a fair amount of joy. So I think being open to all sorts of ways of we only have this life and lots of ways of defining success. And a lot of it relates to are you doing something important, making a difference in people's lives and are you enjoying yourself? Are you feeling, do you have good friends and family with mentees? I think that a lot of it is the same thing. Trying to come into a mentoring relationship without a mold of what success looks like or being too egotistical as success looks like. Look at me and how well I've done. That should be your model. And I've mentored a lot of people to do a whole variety of things over the years. And it's one of the more gratifying things I get to do to see that their success might be working in AI outside of medicine these days. And it's might be being a primary care doctor and it might be being a cardiologist. And there are all sorts of ways of spelling success.
Interviewer/Host
Awesome. And then, Bob, last question for you is you're on the podcast Passion struck. What does it mean to you to live a passion struck life?
Dr. Robert Wachter
Well, I love your emphasis on flourishing. I think that's a word that I don't think people use very much until the last several years. And I think you've partly brought it to public attention. To me, that is, I'm now 68 years old. I try to look at every day, am I making a difference in people's lives? Am I being a good husband, a good father, a good grandfather? Am I doing the things I need to do in work, but also still playing the piano and playing golf and doing the things that give me joy. And on a good day, I feel like I've gotten that balance right. But every day you've got to recalibrate because the world changes around you.
Interviewer/Host
Thank you for that. And Bob, thank you so much for being here. Where's the best place that our listeners can go to learn more about you?
Dr. Robert Wachter
Well, I have a substack now where I'm mostly talking about AI and healthcare. I have websites and things where they can find me. I'm on Twitter, although not very much these days. And my book A Giant Leap will be out, I think, when we post this. So I hope people have a chance to take a look if they're interested in the topic.
Interviewer/Host
Awesome. Well, Bob, thank you so much for joining us today. It was an honor to have you.
Dr. Robert Wachter
Thank you, John. It was a real pleasure.
John Miles
That brings us to the end of Today's conversation with Dr. Bob Wachter. What struck me most is how nuanced this moment really is. AI is not simply replacing humans, it's reshaping the roles humans play in medicine. That could mean less time spent on paperwork and bureaucracy and more time focused on empathy, judgment and human connection. But it also raises profound questions such as who do we trust? How do we guard against bias and misinformation? And what does care look like when machines become part of the decision making process? Dr. Wachter reminds us that technology alone doesn't determine outcomes. Human values, oversight and intention still matter enormously. And that leads directly into our next conversation. On Thursday, I'm joined by Bianca d', Alessio, entrepreneur, real estate leader and founder of one of the fastest growing luxury sales teams in the country. While today's episode explored how systems are changing from the outside in, Bianca's story focuses on transformation from the inside out. We discuss ambition, identity, leadership and what it takes to redefine success on your own terms. It's a powerful continuation of the Life beyond the Script series.
Bianca d'Alessio
I think the most important thing is realizing that we all only have one life to live and we have to live it for ourselves. And you will exhaust yourself so much trying to live life for everyone else. And that's so much easier said than done because I've been on that hamster wheel many times. But I think that's why for me, a constant recalibration of understanding, like what are my goals? Why am I doing this? Is this serving me? Is this relationship serving me and helping me become the best person I want to be.
John Miles
If this episode resonated with you, share it with someone who cares about the future of healthcare. Leave a 5 star rating review on Apple Podcasts or Spotify and explore more reflections from this series@theignitedlife.net until next time. Remember, the future isn't something that simply arrives. It's something we shape through the choices we make, the systems we build and the values we refuse to abandon. I'm John Miles and you've been passion struck.
Whimsound Advertiser
Bring incredible sound into every corner of your home with the new Whimsound smart speaker. Get high resolution Audio with a 1.8-inch touchscreen, smart control and modern design in one powerful speaker for just $299. From Quiet Mornings to lively family gatherings, Whim sound makes every moment sleep sound better and feel better too. Create a home filled with sound you love. Ready to upgrade your sound? Shop now at Amazon and search Whimsound. That's W I I m S o u n D.
Passion Struck with John R. Miles
EP 742: "How AI is Transforming Healthcare and Restoring Humanity"
Guest: Dr. Robert Wachter
Release Date: March 17, 2026
John R. Miles welcomes Dr. Robert Wachter, Chair of Medicine at UCSF and thought leader in healthcare innovation, to explore how artificial intelligence is reshaping the future of healthcare. They discuss Dr. Wachter’s new book A Giant Leap, the opportunities and concerns around AI, and how technology might not only transform systems but actually restore human connection and meaning in medicine. The conversation covers AI’s promises and pitfalls, practical examples, the roots of physician burnout, democratization of healthcare, and what truly matters for both doctors and patients in a rapidly changing landscape.
On AI’s Necessity in Healthcare:
“I can't find a doctor, a nurse, or a patient who says the system is just perfect… nobody thinks that. And I don't think we can deliver on what patients need unless we have the assistance of a technology like AI.” – Dr. Robert Wachter (00:02, 09:57)
On Past Tech Failures:
“The problem wasn’t the technology… [it] does not achieve what you had hoped for… [because] managers bring the technology in and don’t do the hard work of transforming the way work is done.” – Dr. Wachter (17:13)
On Bias and Misinformation:
“We can hurt somebody one patient at a time. If the AI gets it wrong… a chatbot doing mental health that's really bad could hurt a million patients. So the scaling I think is really a problematic issue.” – Dr. Wachter (38:29)
On Empathy and AI:
“We used to think empathy—yet these tools are capable of faking empathy pretty darn well… for the foreseeable future, if I have a new diagnosis… I would want to know that from a human. And I want that interaction.” – Dr. Wachter (48:19)
On Human Flourishing and Success:
“Are you doing something important, making a difference in people's lives and are you enjoying yourself? …There are all sorts of ways of spelling success.” – Dr. Wachter (50:25)
On Human in the Loop:
“For the foreseeable future we're going to need what's called the human in the loop… because the AI is not going to be perfect and we're not going to completely trust it.” – Dr. Wachter (34:04)
On Technology’s Limitations:
“Technology alone doesn't determine outcomes. Human values, oversight and intention still matter enormously.” – John Miles (paraphrased, 53:13)
The episode strikes a thoughtful, authentic, and conversational tone—balancing deep expertise with accessibility and moments of lightness. Both host and guest are open about uncertainties, failures, and the human element within technological systems. There's optimism about AI in healthcare, tempered by realism about past mistakes and future risks.
This episode offers a balanced, insightful roadmap for anyone concerned with the future of healthcare and human flourishing in a world increasingly shaped by AI. Dr. Wachter’s perspective—born from decades on the frontlines of medicine and system-level reform—is one of hope, cautious optimism, and belief in the enduring importance of human connection, even as technology rapidly evolves.