Transcript
A (0:00)
You could sort of have a situation where the virtual beings almost like dominate the humans in every single axis of moral value. And now suddenly it starts to look like criminally decadent to be spending kilometer east land on the legacy humans. It's hard to draw a hard line between having a permanent income and being like some sort of parasite. And that's going to be the cultural battle. And it's going to be really easy to say that humans are parasites once we're not providing value to the larger growth engines. Yes, there will be cool, interesting stuff, stuff happening in the future if we allow competition to run, but we just probably won't be meaningfully part of that. A lot of people are basically going to say my only hope is to sort of be the first one to bow down to the new overlords and embrace this new culture. Sometimes people ask like, oh, aren't corporations super intelligences? Why shouldn't we fear them? And the answer is because it's made of people, so it needs us. There's this idea of the singularity which I feel like has been very destructive because it kind of is like an excuse to turn off your brain and to not model the future and to say, yes, things will keep changing faster and faster until we can't say anything about it. Let me say it loud and clear here, is that, yeah, I think that the post AGI world is just going to be extremely alien and so different that if we could avoid crossing that threshold, I think we should.
B (1:05)
David, welcome to the show. Thanks for being here.
A (1:07)
Thank you for having me, Gus.
B (1:09)
Great. Do you want to introduce yourself?
A (1:11)
Sure. So my name is David Duveno. I'm an associate professor of computer science and statistics at the University of Toronto. I've been working on probabilistic deep learning for a number of years. And then I guess in the last few years I decided to use my freedom to try to focus on the problems that seemed most neglected and intractable, which has led me to first work on more technical AI safety. So I was a team lead at Anthropic for a year and a half starting in 2023, working on sabotage evaluations. And then I felt like there's this sort of me and other people were noticing there's an even bigger missing part of the alignment problem, which is how do we align our entire civilization, which of course is even more intractable and even hard to describe this problem. So that's what I've been thinking about lately.
B (1:55)
You have this fantastic paper on gradual disempowerment. So let's start there. Can you explain how gradual disempowerment would be different from an AI takeover?
