Transcript
David Bau (0:00)
Historically, as an engineer, as a computer scientist, it's really been our responsibility to understand the systems that we make, to make sure that they are doing what we want, that they operate correctly. And I feel that the new discipline of machine learning, because it's become so important to just accept these black boxes and use them even though we don't understand them, is leading to really an unhealthy turn in the practice of engineering in computer science. We're training a whole new generation of computer scientists to be comfortable with this idea that they shouldn't really look inside these complicated black boxes.
Jasia Monk (0:46)
And now the good fight with Jasia Monk.
Podcast Narrator (0:55)
There is, at the moment, for good reason, a lot of debate about artificial intelligence. How capable are current models? Are we approaching something like artificial general intelligence? And how big are the impacts on the world going to be? Is this, for example, going to lead to a massive loss for jobs, not
Jasia Monk (1:15)
only for drivers and other kinds of
Podcast Narrator (1:17)
blue collar, but professions, but also perhaps for a lot of white collar professionals?
Jasia Monk (1:23)
Well, one thing that strikes me about
Podcast Narrator (1:25)
this debate is that it's often held without real knowledge of the underlying technology. And so I was really intrigued to learn more about the nuts and bolts of how AI actually works. And when I was recently at a conference about artificial intelligence at Harvard, I had the good fortune of running into David Bao. David was at Google for a long time, is now a computer scientist at Northeastern University, and he is really good at explaining what a neural network is, how it is that you train an AI model, and what that means for how to think about this technology. So I invited him on the podcast and got him to give me and you a 101 introduction into how AI models actually work.
Jasia Monk (2:25)
David Bau, welcome to the podcast.
David Bau (2:27)
Thank you for having me.
Jasia Monk (2:28)
I really look forward to this conversation. We met recently at a workshop about artificial intelligence at Harvard, and I thought that in a conversation we had, you had helped me better to understand the nature and the architecture and the technology of artificial intelligence than anybody had before. So I think, Ford, I would love to talk to you about this on the podcast. Fundamentally, how do current AI models work? When we say they are LLMs, large language models, what does that mean? And how does that distinguish this kind of form of artificial intelligence from other forms that we've historically used?
