
Loading summary
Alok Jha
Hello. This episode of Babbage is available to listen to for free. But if you want to listen every week, you'll need to be an Economist subscriber. And good news. We're running a Black Friday sale between Thursday the 23rd and Monday the 27th of November. You can sign up to Economist Podcast plus for half price. For just $2 or pounds or euros a month, you'll get access to all of our award winning podcasts. For full details, search online for economist podcast plus.
Fei Fei Li
My dad works in B2B marketing. He came by my school for Career Day and said he was a big roas man. Then he told everyone how much he loved calculating his return on ad spend. My friends still laugh at me to this day. Not everyone gets B2B, but with LinkedIn you'll be able to reach people who do get $100 credit on your next ad campaign. Go to LinkedIn.com results to claim your credit. That's LinkedIn.com results. Terms and conditions apply. LinkedIn the place to be to be.
Alok Jha
The economist ChatGPT. Maybe you've heard of it. If you haven't, then get ready. Sometimes it can feel like artificial intelligence just appeared out of the blue a year ago because this promises to be the viral sensation that could completely reset how we do things. For most people, ChatGPT was a shock. This algorithm could produce sentences almost like a human. It was uncanny and perhaps a bit scary. Our guest today, though, well, she saw it all coming. In fact, Fei Fei Li laid the critical groundwork for the technology that underpins large language models and image creation models, and which has accelerated progress in AI in the past decade. Fei Fei Li is a computer scientist at Stanford University and the founding co director of the university's Institute for Human Centered Artificial Intelligence. She made her name in the field as a pioneer in developing ways for computers to see and and recognize objects. Computer vision is nowadays so good that it's embedded in everything from border checkpoints to driverless cars, and it's even involved in helping us make better video calls. Computers have become even better than people at recognizing human faces, which has also led to a host of ethical questions on how this technology should be used in public. In developing the field of computer vision, though, Fei Fei Li also sparked something much bigger the use of enormous data sets as a way to train AI models. In the mid 2010s that led to the revolution in deep learning. And deep learning is at the heart of the biggest AI models today. In our conversation AI and in her new book, the Worlds I See, Fei Fei Li talks about the challenges of her early work in computer vision, how it led to chatbots and the large language models we see today, and how we should all be thinking about the opportunities and the risks of the next generation of AI.
Fei Fei Li
I worry about it a lot, especially as a technologist who actually understand what this technology is. And I think we have to take responsibility to govern it as well as to mitigate these risks.
Alok Jha
I'm Alok Cha and this is Babish from the Economist. Today, Fei Fei Li, a pioneer of AI on its future. Fei Fei Li, thank you for joining me.
Fei Fei Li
Thank you Alec for inviting me.
Alok Jha
Fei Fei, I think it's worth talking about the current age of AI that everyone is fascinated by at the moment. And I'd love to hear your thoughts on generative AIs, the type of AIs that power things like ChatGPT and Bard from Google and others, they're all based on deep learning neural networks of the type that you know, were really revolutionized by your work. And you must have seen the language models getting better and better as well. What was your reaction to the public release of ChatGPT in 2022? The world seemed to go crazy about it, but I wonder what you thought.
Fei Fei Li
My world went crazy as well. So as a technologist, my colleagues and friends, very deeply in the weeds of AI. We probably saw this a little earlier than the world. When GPT2 was released, we especially my colleagues here at Stanford, were recognizing that this way of training transformer models with huge amount of diverse data will give us the kind of powerful algorithm like ChatGPT eventually in ways we've never seen before. Of course, we didn't Predict the precise November 2022 would be the release. We didn't predict that, but we actually anticipated that by creating a center in my institute called center for Foundation Models to look at the different aspect of these algorithms. So from that point of view we did anticipate, but. But yet the public awakening the inflection moment of the entire world in response to this technology was still a huge moment. It accelerated our work and concerns about how we talk about this, how we educate our students, to our policymakers, to our business executives, to our journalists and reporter friends. It accelerated our work in engaging with from Washington D.C. to Brussels in terms of the policy and regulatory framework surrounding this technology. And really completely serendipitously, just a few months later, my book was shipped to the publisher. And I think it really was a moment that reinforced my own belief that no matter how powerful the technology Is like GPT, like technology. It has to be grounded in a human centered approach.
Alok Jha
Fei, Fei, you're well known for starting your career in the attempts to try and build computers that can recognize objects and scenes. This thing of computer vision. Now, just before we go into the technological aspects of that, what do you think it means for a machine to see something and recognize something?
Fei Fei Li
Yeah. I want to invite you to imagine what it means to see, period. 540 million years ago, the simple animals living under the water don't see. They don't have something like ice, they just float around. They don't get to feel the world, they don't sense the world. The world is in kind of a darkness that they don't even know it's darkness. And then around 500 million years ago, they start to develop a little pinhole almost in their region where they would call head and start to collect lights. And that moment must be such a fascinating or even poetic philosophical moment of animal life. Right. When you get to receive information from the world. And some zoologists actually hypothesize that the ability to see as one of the oldest sensory capabilities of animals was what set off the evolutionary arms race of animal speciation. Because once you can see, you can see dinner, you can become somebody else's dinner.
Alok Jha
You can avoid being dinner as well.
Fei Fei Li
Exactly. You start hide, you start to run, you start to evolve. And the course of intelligence, the nervous system, started. So seeing is a cornerstone of intelligence. It's part of how we understand the world, part of how we survive. And by the time you get to humans, we are such visual animals, we use seeing as a way of learning, as a way of interaction, as a way of making things happen, as a way of socialization, as way of communication. So this is why vision fascinates me. Cracking the fundamental questions of vision is cracking the questions of intelligence. And that's why I specialized in computer vision.
Alok Jha
And so intelligent systems, in your telling, have to be able to see because that's such an important part of how they behave or how they can survive.
Fei Fei Li
Essentially, by and large. Yeah.
Alok Jha
So what does that mean in terms of a machine then? How do you build a system, a computer system that can, quote, see?
Fei Fei Li
Yeah. Well, one thing we should recognize is that cameras take pictures. And that's a very physical process of allowing photons to land on some kind of sensors. Used to be films, now it's digital to register color, to register illumination or grayscale lighting. And that is only a very small part of seeing. And we have that with our own Eyes, right? We have retina that gets stimulated when we see lights and of different color. But real seeing happens in the computation of making sense of the sensory data. Otherwise they're just numbers or they're just electrical poles coming from the retina. But now I'm talking to you. You know, the different colors and shades and pixels coming from you make me retain recognize. I'm talking to Alec, I'm talking to a person who's wearing a blue shirt. When you nod, I get very happy. And that is making sense and understanding of what we are seeing. And that is a non trivial feat because there are so many things, there are so many possibilities. Mathematically, you know, if you want to assemble pixels in different composition, it's infinite. Yet we actually see the world in such richness and beyond understanding. We actually interact with the world. When you make a sandwich this morning, you have to do so many things with your visual intelligence from, you know, opening the fridge, understanding where the ingredients are, bringing them out. You, you have to use your eye to guide your hand. If you close your eye, your hands will not be able to grab that piece of tomato or bread. And then you have to use a knife to spread the butter and it just goes on and on. Right?
Alok Jha
So you've been working on this for some time and I just wonder if you could just give me a brief history of how computers started to be able to see.
Fei Fei Li
We began as a field in the 60s and 70s, trying to just make sense of very simple lines or simple geometric shapes. Yet as we move into the more ambitious realm of computer vision, we start to look at tackling the problem of recognizing objects. And that was really my PhD study. I remember the first data set I worked with had four different objects. Airplane, leopard face and the car. You would ask me, how did you pick these four objects?
Alok Jha
How did you pick these four objects?
Fei Fei Li
First of all, I didn't. My PhD advisor said, Fei Fei, here's someone else's data set. And they made these four objects. And it's already hard enough to get a few dozen or a couple of hundred pictures. That's back 2000. It's really hard. Digital cameras were just about to take off. There was no smartphone. The Internet was barely born. Throughout my entire PhD it was working on these very small dataset. And not only my PhD was small dataset, my entire field was working on this small data set.
Alok Jha
In 2006, Fei Fei Li began to create a database of images, each one carefully labeled to describe what was in the picture. ImageNet, as the database was called. Was part of a project to develop AI algorithms that could be trained to see the world and recognize objects. The purpose of ImageNet was to widen the set of reference images for the computer science community from four to many thousands.
Fei Fei Li
ImageNet gets you to train algorithms to recognize tens of thousands of different kind of objects. It also is not only a data set that is ginormous 15 million images across 22,000 classes of object. It's also a mind shift for how we train AI systems. ImageNet was the turning point of AI's history. That's recognizing how critical it is to use big data.
Alok Jha
Every year, the ImageNet team would run a competition to see which algorithms were best at recognizing objects they'd never seen before. In 2012, when the competition was being held in Italy, Geoff Hinton, an AI researcher from the University of Toronto, entered his algorithm, Alexnet. It was so good that Fei Fei Li rushed from her home in California to watch the results.
Fei Fei Li
Their algorithm was actually a classic class of algorithm. It was called neural network, or in this case, convolutional neural network. It was invented in the late 1980s, but they combined that with Imagenet with two GPUs and made huge progress in terms of reducing the error of the tests and winning the competition. And that was when I recognized that it was the watershed moment for AI's evolution.
Alok Jha
Alexnet used a type of AI built on many layers of neural networks.
Fei Fei Li
Every layer is consisted of little nodes that function like neurons, similar to neurons, mimics neurons. I would not say it replicates neuron. You begin with an image. The first layer of neurons take the values of the colors of each pixels of the image, and then all these neurons compute. And then you keep computing. And Alexnet has about seven layers. After seven layers, it says, oh, after all this, I see cat or I see a microwave. And that's really what this neural network does. But what is really hard is how do you teach this neural network to grab information that's important? For cats vs microwaves vs chairs?
Alok Jha
The process was called deep learning.
Fei Fei Li
So at the beginning, what you do is you feed this neural network, in this case, millions of images. Every image has a label, right? Some are cats, some are microwaves, some are chairs. And then, you know, you have to learn to differentiate cats from microwave from chairs. And if you're wrong, you know you're wrong because you know the ground truth label. And the algorithm sends back a signal and say you have to correct this parameter that gets you this value, and you do it many, many, many, many millions of times. And eventually you learn a network with enough mathematical values that do a decent job in capturing cats information or microwave information, and then you use that to go about recognizing object in the world.
Alok Jha
Deep learning would go on to become the basis for the AI. Models have taken the world by storm in the past year. Now, you talked about your history in starting with your colleagues, this idea of how to recognize objects and it zooms all the way forward to, you know, smartphones that can recognize faces. We see supermarkets that don't need checkouts, robots can be guided around factories. It's just really permeated so many parts of life and industry. What do you think the most exciting use for computer vision is today?
Fei Fei Li
Yeah, Alec, that's a great question. Just think about humans and animals. We use vision for everything. I genuinely think vision will empower and enable people and machines to do a lot of things eventually. I think if we look for what we call embodied AI, you know, robots and doing machines, I think vision is their cornerstone. Without making cars to see, we won't have self driving car. Without making cameras to see, we won't be able to guide, say surgeries without sensing in large scale, we won't be able to map out the biodiversity of our planet. We won't be able to go under the deep ocean for exploration, we won't be able to, you know, train algorithms to tell cells with pathology from healthy cells. So we really have a boundless possibility using vision. I myself am working in robots in robotic learning. I'm very excited by that. I'm also working in healthcare where we use camera sensors as guardian angels to help doctors and caretakers to watch our patients for safety and other clinically relevant conditions.
Alok Jha
Can you give me a timescale for the embodied AIs that we are talking about, when robotics and everything will really start to integrate?
Fei Fei Li
Yeah. You know Alec, I've been asked this question a lot and frankly I don't even think my own answers are necessarily consistent. It shows how hard it is to predict the future.
Alok Jha
That's moving fast, right?
Fei Fei Li
Yeah, it's moving fast and I absolutely believe it in my lifetime because that's why I'm working on it. I hope I live long.
Alok Jha
Well, I'm sure you will. It's also possible that machine vision is going to have lots of perhaps worrying uses and in fact they're already here. Things like unwanted facial recognition in public or, you know, using them to guide weapons, for example. I wonder, as one of the people who has made all of this possible, how much concern do you have about these sorts of negative uses if you like. Do you think about them at all?
Fei Fei Li
Absolutely. That's why I wrote the book. That's why for so many years I've been advocating for human centered AI. Our relationship with AI is like our relationship with all tools. It's a double edged sword. Humanity. It's in our DNA to be innovative. We want to be innovative because we want to make life work better. Right from the the moment we discover fire or use that first stone axe, we are inventing tools and now we're in the intelligent machine age. But our fundamental relationship with tools haven't changed is that we can use it for good and we can use it to do harm. Some harms are intentional and some harms are unintended, but still equally bad. So I worry about it a lot, especially as a technologist who actually understand what this technology is. I see how it can be wrongly used, how it can be wrongly communicated, how can be harmful to people. And I think we have to take responsibility to govern it as well as to mitigate these risks.
Alok Jha
We'll look at how to solve those challenges in AI with Fei Fei Liquid next. But first, just a reminder that this is a free episode of Babbage. To listen to us every week, you'll need to become a subscriber. If you already have a subscription to the Economist, you just need to link your podcast app. There's a link to a video explaining how to do it in the show Notes. If you're not yet a subscriber, you can make the most of our Black Friday deal and sign up to Economist Podcast Plus. For those of you that are subscribers, we've posted a little bonus for you in the Babbage Feedback. Every couple of weeks, we'll bring you one article from the printed edition of the Economist. Read aloud the story we've picked for you this week looks at how to improve scientific research. We argue that more experimentation is needed in funding mechanisms to supercharge science. Find that article in your podcast.
Fei Fei Li
Apparently, everyone knows that it doesn't truly feel like Valentine's without the flowers. So whatever you do, don't forget the flowers. Beat the Valentine's rush and order early at 1800-flowers.com Right now, when you order early from 1-800-flowers, you can get up to 40% off gorgeous bouquets and one of a kind arrangements that are guaranteed to wow. Don't put this off. Delivery dates are limited, so get up to 40% off today at 1-800-flowers.com acastra. That's 1-800-flowers.Com acast.
Alok Jha
Today, we're speaking to a pioneer of artificial intelligence, Fei Fei Li. Earlier in our conversation, she explained the need for transparency and better communication around new generative AI technologies. As we've explored before on Babbage, generative AI raises a lot of challenges for society. Earlier this month, the British government hosted a summit on AI safety at Bletchley park just outside London. This was where the country's mathematicians broke the German Enigma code during the Second World War. 28 countries, including the US and China, have signed up to the Bletchley Declaration, an attempt to kickstart a global approach to manage the very latest technology. Amid warnings from testing tech giants that it presents the greatest existential threat to humanity. And in America, President Joe Biden has issued an executive order that sets out new standards and regulations for how AI should be built in the future. Signing an executive order that would require.
Fei Fei Li
Developers working on the most powerful AI models to share their safety test results with the government.
Alok Jha
There are a wide range of worries and risks about generative AI and computer vision. How do people like Fei Fei Li, who brought these technologies into the world, think about their ethical challenges? Let's rejoin my conversation with her.
Fei Fei Li
So, Alec, look, I think we all have responsibility. It's such a complex and powerful technology. It's impossible that there is one solution. To me, education and communication is a critical part of it. I don't think we have been doing enough public education or enough good public education. I don't think going after hyperbole is a helpful way of public education. This is a very nuanced technology. But we really do need to educate the public much, much better on also the risks, from disinformation to bias, to privacy concerns, to jobs changes. But that's one part of responsibility. We also have responsibility as developers and inventors of this technology. How do we consider the ethical values? How do we consider responsible and transparent development? And then we have responsibility in deploying this technology for business that's creating products and services. What do we bring to our consumers, to our community? And last but not the least, civil society and governments have the responsibility for governing, for guardrailing, for creating the right framework or ecosystem to use this to do good, but avoid catastrophic risks that we might be dealing with because of the power of this technology.
Alok Jha
What are your concerns about technologies like ChatGPT? There are lots and lots of things that they will do that will benefit and we can talk about those. But when you see these technologies and the things they're capable of and then the conversation around them, what Are the things that sort of worry you about their capabilities.
Fei Fei Li
Yeah. So like you said, I do want to give a very nuanced approach. It has a lot of powerful positive empowerment and applications and possibilities for people. But you're asking about the concerns. We talk a little bit about bias. Human civilization has been biased. And now we have powerful machines that can propagate our historical bias. And that's all good. That would hurt people, especially people from underrepresented, underserved, undervoiced communities. Whether it's automated machine hiring algorithm or a skin disease detector by AI or bank loan assessment, you can see that the bias could potentially propagate. And that's just one bucket of catastrophic risk. There's another bucket, for example, disinformation. I'm deeply, deeply concerned. Of course, humans have been propagating disinformation since the beginning of civilization. But technology lowers the entrance bar, right? So you can now scale up so easily. You can now create fake videos, fake audios, fake podcasts. And that would be detrimental to a society that is built upon trust. There is a civil contrast of trust of information, and that would be detrimental. And how do we combat that? And then there's jobs. Clearly there will be job disruption. There will be also job empowerment. You know, this technology can increase productivity, but in the meantime, productivity doesn't necessarily translate to a fair distribution of prosperity and wealth in workforce and our society. Our history, everywhere globally has seen this movie over and over again. So how do we deal with this wave of technology shift? We can go on and on.
Alok Jha
We asked some of our listeners to send in some questions about things that they were worried about. When it comes to large language models and the capabilities, there seems to be a lot of concern about hallucinations. I mean, these are the things where chatbots and AR models will just invent information to answer the questions. And one of our listeners, Andrew, a microbiologist in California, he asks, can large language models ever be fixed to mitigate the problem of hallucinations? Is it something that's just a technology problem, or is there something deeper going on here?
Fei Fei Li
That's a great question, Andrew. First of all, if you look at the progression of, say, OpenAI and other large language models, this problem is being combated. You know, every release of, say, the GPT products has been better with hallucination. And a lot of this is what they call the reinforcement learning from human feedback training or post training that is trying to combat hallucination. But is this a pure technical issue to solve? I believe it's more Nuanced, for example, it really does depend on data. Right? So if you feed, let's say I say the sun is green and I feed the algorithm a lot of data about the sun is green, well, the algorithm is going to learn the sun is green. And that is not technically hallucination because it is trained on that. So starting with data integrity, we have potential issues. Of course, because of the way the model is made, it might always hallucinate just out of the patterns it has learned. So we do also have to have a layer of responsible use. How do we communicate with users? How do we actually provide sources of information? These are important questions from product and service point of view.
Alok Jha
I wonder what the next sort of iterations of generative AI look like. Are images and video going to be useful data sources for things like language models in the future? Are they going to be the thing that transforms it next time?
Fei Fei Li
Absolutely. Actually, if you look at the most recent releases including GPT4V, it's already beginning to be multimodal. And a lot of this training have used images and videos and there has been really interesting work in these diffusion models in generative AI images. I mean, you know, you type a sentence like put me in a sushi restaurant and you actually can get pictures of sushi restaurants. So we're already seeing that multimodality.
Alok Jha
Slightly terrifying actually, in a way, yes. For your misinformation point, yes.
Fei Fei Li
I'm not going to deny that actually both language and now images and videos, as we talked about it, are definitely fertile grounds for mis and disinformation. But from technology point of view, multimodality is absolutely the future. And then the ability to translate the physical world into a digital one, and vice versa, is going to bridge the work of robotic learning with today's more software layer AI. So there's a lot to come from technology point of view. But you're right though, every one of them comes with its own risk and potential misuse. That is the concerning part.
Alok Jha
Just one more point on this, which is generative AI from your point of view has so many advantages and there's so much to go. But you've pointed out the things like misinformation, disinformation, bias, these things are real concerns to people right now and should be regulated against and people should be aware of them. But so much of the conversation around generative AI and the future of AI is around these so called existential risks, you know, that's going to destroy the world somehow or it's going to turn us all into paperclips or whatever else. And even though there's no evidential pathway towards those things, people seem to be very concerned about them, including governments. And there have been AI safety summits in the uk. President Biden's issued his executive orders. The heads of these companies go to these government agencies all the time to talk about these things. I wonder what you think about this. Did you think that people focus too much on these catastrophic and existential things just because they're kind of terrifying to the detriment of worrying about the real risks that people are facing now?
Fei Fei Li
I had this public debate with Professor Geoff Hinton, who's a personal hero of mine as a technologist and he definitely is one of the vocal people talking about the potential existential possibilities. And I actually respect that. In the meantime, a respectful philosophical pondering about the future of this technology doesn't substitute pressing urgent social catastrophic issues. We need to take responsibilities now to combat and we have been saying things like bias, jobs, weaponization, disinformation. And there is another social issue that is very urgent and pressing is the lack of investment in public sector development of this technology. And public sector has always been a more trusted source of honest, transparent assessment and evaluation of this technology. And not a single university in the US can train a CHATGPT model today. And with this kind of deprivation of public sector, we will have long term crisis in terms of training the right talents in terms of depriving us from scientific discovery and depriving us from evaluation and assessment of this technology. So I actually think it's critically important, especially for governments to talk more about the immediate pressing social issues as well as creating ecosystem that would benefit the public. For example, investment in public sector or proper, hopefully not overly done, not under proper regulatory framework, rather than for the governments to engage in long term philosophical existential crisis rhetoric. With all due respect to policymakers, they are not trained to do that. We've got a lot of scholars and true deep thinkers that are probably more trained to do that. But the policymakers are tasked to help the public make the right policies for their regions, countries and world.
Alok Jha
And when you say investment by public sector bodies, do you mean trying to have models or clouds of computing that basically are available to public institutions, governments, universities, et cetera, so that not everything happens within companies.
Fei Fei Li
Yes. So President Biden issued an executive order from the White House just days before the UK AI summit talking about the kind of government measures for AI. Some of them were regulatory measures, but some of them were public investment. Most notably is a pilot Program for the National AI Research Cloud that we call nair National AI Research Resource. I was actually personally involved with my Stanford colleagues in leading this effort. Four years ago, 2019. We wrote an open letter about the need for Nair in 2020. And we worked on an early bill called the NAIR Task Force Bill under Trump administration that made it happen. And then I served as one of the 12 task force members for the White House to plan for nair. And now we are also part of the leading voices from public sector to endorse the Create AI act bipartisan bill being considered in both US Congress and Senate.
Alok Jha
Fei, Fei, I can't let you go without talking to you about the future of AI and what we should be thinking about and looking out for in the coming years. And I suppose one sort of usual journalistic way of dealing with this is to ask you about artificial general intelligence. This idea of a machine that can do all the things that a human brain can do, should we ever want to build on those things? You know, generative AIs have tricked us into thinking that there's something intelligent going on underneath the hood, when in fact it's just statistical meanings of language and so on. You know, they've captured the public imagination. But do you think that these are the route to the mythical artificial general intelligence, or do you think we're going to need more than that?
Fei Fei Li
So, Alec, from a technological advanced point of view, I think we are creating very powerful machines. And as we become more multimodal, as we create embodied AI, robots and agents, we're creating machines that will be able to do a lot of what humans can do. I can't wait for a machine to help me to clean toilets and bathrooms. But in the meantime, human intelligence and human lives is actually very, very high dimensional, very layered, right? The kind of feelings we have for each other, the kind of creativity that sparkled our artists, our scientists, our writers, musicians, empathy, compassion. I have no way of even grasping as a technologist how to, quote, unquote, model that, how to write an algorithm to do that. And I think that will be an important question for each individual. If we create machines that can do some of the tasks humans do today, does that mean it's replacing the essence of being human? I actually don't think so. Just like, you know, humans move, but we've created, from bicycles to cars to rockets, it didn't take away our agency of moving, we created powerful tools that help us to move or even move us. So I think these are important questions to ponder. I'm not downplaying the important both side the risks and possibilities of AGI. I think it's really worth talking about. I just want to make sure that we see ourselves with agency, with multidimensions, with depth and nuances. That is how we define who we are.
Alok Jha
How about a sort of extreme question? Do you think that AIs could ever become sentient? Are they going to become living in some way or conscious in the way that you have in science fiction all the time?
Fei Fei Li
This is not an extreme question. This is what I say that I respect this kind of discussion. I respect this kind of scholarly work. There's definitely already some level of cognitive understanding of language, let's say of these algorithms. But are they like sentient or aware or conscious in the way of humans? Absolutely not yet. Is there a way to create silicon based computational algorithm that will get to that level of self awareness? I remain open minded. I think there's a long way to go. But I think it's a worthy philosophical, scholarly study because we'll always be curious about it. Humans are curious. Humans go after these kind of problems. And it's also important to study this because like we said, only when we know something, we know how to evaluate, assess and govern it.
Alok Jha
I think that sounds like a good place to end. You seem like a very optimistic person when it comes to AI, whilst always also acknowledging the troublesome aspects, which I think is probably the best way to be about it. So BEI Fear Lee, thank you very much for your time. Thank you Alok, and thank you for listening. If you want to keep up to date with the very latest goings on in the world of AI this week, then head over to our sister podcast, Money Talks. My colleagues there will be looking at the business side of artificial intelligence and what the recent turmoil at luxury OpenAI means for the future of the industry. Money Talks comes out on Thursdays. Finally, don't Forget that as COP28 rapidly approaches, we'd like to know if you've got any climate related questions. Get in touch with us by emailing podcastseconomist.com and make sure you put Babbage in the subject line. You might even get your questions answered on a future show. This week Babbage was produced by Jason Hoskin and mixed by Johnny Allen. The executive producer is Marguerite Howell. I'm Alok Jha and in London. This is the Economist.
Fei Fei Li
My dad works in B2B marketing. He came by my school for Career day and said he was a big roas man. Then he told everyone how much he loved calculating his return on ad spend. My friends still laugh at me to this day. Not everyone gets B2B, but with LinkedIn you'll be able to reach people who do. Get a $100 credit on your next ad campaign. Go to LinkedIn.com results to claim your credit. That's LinkedIn.com results. Terms and conditions apply. LinkedIn the place to be to be.
Babbage: Fei-Fei Li on How to Really Think About the Future of AI
Published on November 22, 2023, by The Economist
In this insightful episode of Babbage, host Alok Jha engages in a compelling conversation with Fei-Fei Li, a renowned computer scientist at Stanford University and a pioneer in the field of artificial intelligence (AI). Li, the founding co-director of Stanford's Institute for Human-Centered Artificial Intelligence, delves into the evolution of AI, the profound impact of computer vision, and the ethical considerations that accompany the rapid advancement of generative AI technologies like ChatGPT.
Fei-Fei Li begins by reflecting on the public's reaction to the release of ChatGPT in 2022. While the general populace was taken aback by the sophisticated capabilities of the AI, Li and her colleagues had anticipated this breakthrough due to their foundational work in large language models and image creation algorithms.
Fei-Fei Li [06:27]: "But yet the public awakening the inflection moment of the entire world in response to this technology was still a huge moment."
Li emphasizes the importance of grounding AI advancements in a human-centered approach, ensuring that technological progress aligns with societal values and needs.
The conversation shifts to the concept of computer vision, a field where Li has made significant strides. She articulates the profound implications of enabling machines to "see" and recognize objects, drawing parallels to the evolutionary milestones of vision in biological organisms.
Fei-Fei Li [07:59]: "Seeing is a cornerstone of intelligence. It's part of how we understand the world, part of how we survive."
Li underscores that vision is not merely about capturing images but involves complex computational processes that interpret and make sense of sensory data, mirroring human visual intelligence.
A pivotal moment in AI history, according to Li, was the creation of ImageNet in 2006. This extensive database of labeled images marked a significant shift from limited datasets to expansive, diverse collections that fueled the deep learning revolution.
Fei-Fei Li [12:55]: "ImageNet was the turning point of AI's history. That's recognizing how critical it is to use big data."
ImageNet enabled algorithms to recognize tens of thousands of objects, drastically improving the accuracy and capabilities of computer vision systems. The annual ImageNet competition further accelerated advancements, highlighting breakthroughs like Geoff Hinton's AlexNet in 2012, which leveraged convolutional neural networks to achieve unprecedented performance.
Fei-Fei Li highlights the myriad applications of computer vision, illustrating its transformative role across various industries. From self-driving cars and robotic surgery to biodiversity mapping and healthcare monitoring, the versatility of computer vision is evident.
Fei-Fei Li [16:54]: "Without making cars to see, we won't have self-driving cars. Without making cameras to see, we won't be able to guide surgeries."
Li is particularly excited about embodied AI and robotic learning, envisioning machines that seamlessly integrate visual intelligence to perform complex tasks and enhance human capabilities.
Addressing the darker side of AI, Li expresses deep concern over issues like bias, disinformation, privacy breaches, and job displacement. She warns that without responsible governance, AI technologies can perpetuate historical biases and exacerbate social inequalities.
Fei-Fei Li [19:07]: "Humanity's fundamental relationship with tools is that we can use it for good and we can use it to do harm."
Li advocates for a multi-faceted approach to mitigating these risks, emphasizing the roles of education, transparent development, responsible deployment, and robust regulatory frameworks.
The discussion broadens to encompass global efforts in AI safety and regulation, including the Bletchley Declaration signed by 28 countries and President Joe Biden's executive orders aimed at establishing new standards for AI development.
Fei-Fei Li [24:58]: "I think we need to take responsibilities now to combat and we have been saying things like bias, jobs, weaponization, disinformation."
Li highlights the importance of public sector investment in AI research and the development of resources like the National AI Research Cloud (NAIR), which aims to democratize access to AI technologies and foster transparent evaluation.
Exploring the horizon of AI, Li discusses the concept of Artificial General Intelligence (AGI)—machines that possess the ability to perform any intellectual task that a human can. While acknowledging the technological advancements leading toward AGI, she remains skeptical about machines achieving human-like consciousness and emotions in the foreseeable future.
Fei-Fei Li [36:11]: "Human intelligence and human lives are actually very, very high dimensional, very layered... I don't think that replaces the essence of being human."
Li emphasizes the unique aspects of human intelligence, such as creativity, empathy, and consciousness, which remain beyond the reach of current AI systems.
When probed about AI's potential for sentience, Li maintains a clear distinction between advanced computational abilities and true consciousness.
Fei-Fei Li [38:18]: "Are they like sentient or aware or conscious in the way of humans? Absolutely not yet."
She advocates for ongoing philosophical and scholarly exploration of these topics to inform responsible AI development and governance.
Fei-Fei Li's dialogue with Alok Jha offers a balanced perspective on the future of AI, blending optimism for its transformative potential with a sober recognition of the ethical and societal challenges it presents. Her advocacy for human-centered AI and proactive governance underscores the necessity of aligning technological advancements with the broader aspirations and values of humanity.
For those interested in diving deeper into the business implications of AI and recent industry developments, The Economist's sister podcast, Money Talks, explores these themes every Thursday. Additionally, as environmental concerns become increasingly pressing, listeners are encouraged to engage with climate-related discussions by submitting questions to Babbage for future episodes.