Transcript
Sebastian (0:00)
We can take an example of how nature evolved intelligence and use evolution instead. When you use a static fixed network that is not changing the weights during its lifetime, if you cut off a leg, it will probably fail because it can't adapt. But these Hebbian networks, they change the weights all the time. It's basically like a continually learning, updating system where you can cut off a leg and oftentimes it will still be able to function, even though it has never seen this kind of variation during training.
Host (0:27)
Let me jump in with a little explanation before we get started. This is a very technical podcast, but one of the more interesting ones that I've recorded in a while, and I want as many people as possible to benefit from it. I'll begin by explaining in simple terms what gradient descent is, which is used in most neural networks today, as opposed to neuroevolution, which is what this podcast is about. An illustration of gradient descent is standing blindfolded on a mountainside with the goal of finding the lowest point in the landscape. That lowest point is the solution. Your distance from it is the error or loss to get to the solution. You try to reduce that error step by step. So you feel around with your foot to find which direction the ground slopes downward. You then take a step in that direction. You repeat that process until every other direction feels uphill. At that point, you've reached a low point called a minima, though not necessarily the lowest point in the whole mountain range, which would be called the global minima. With neuroevolution, imagine a plane flies over the whole mountain range and drops many people with different search strategies in many different places. One wanders to his left, another to his right. One walks in widening circles, another takes big jumps, another small ones. After a while, you see who ended up at the lowest point. You keep the best strategies, make variations of them, combine some of them, maybe the jumping and the walking in a widening circle. And then you send out a new group of people with those strategies. Over time, your people get better and better at finding the lowest point, even though none of them ever knew which way was downhill. That is the difference. You have a better chance of finding the global minima, the lowest point in the entire mountain range. With gradient descent, you improve by following the slope. With neuroevolution, you improve by variation and selection. You try many candidates, score the results, keep the better ones, and make new variants from them. No one has to know which way is downhill to begin.
Interviewer (3:19)
You have this new book out, Neuroevolution, so maybe you can Start by explaining what neuroevolution is in AI.
Sebastian (3:28)
