Transcript
A (0:00)
Any LLM that was trained on pre1915 physics would never have come up with a theory of relativity. Einstein had to sort of reject the Newtonian physics and come up with this space time continuum. He completely rewrote the rules. AGI will be when we are able to create new science, new results, new math. When an AGI comes up with a theory of relativity, it has to go beyond what it has been trained on to come up with new paradigms, new science. That's my definition of AGI.
B (0:31)
Vishal Mistra was trying to fix a broken cricket stats page and accidentally helped Spark One of AI's biggest breakthroughs. On this episode of the A16Z podcast, I talk with Vishal and A16Z's Martin Casado about how that moment led to retrieval, augmentation generation, and how Vishal's formal models explain what large language models can and can't do. We discussed why LLMs might be hitting their limits, what real reasoning looks like, and what it would take to go beyond them. Let's get into it.
C (1:00)
Martin, I know you wanted to have Vishal on. What do you find so remarkable about him and his contributions that inspired this?
D (1:06)
Vishal and I actually have very similar backgrounds. We both come from networking. He's a much more accomplished networking guy than I am.
C (1:11)
That's a high bar given you from.
D (1:13)
And so we actually view the world in an information theoretic way. It is actually part of networking. And with all this AI stuff, there's so much work trying to create models that can help us understand how these LLMs work. And in my experience over the last three years, the ones that have most impacted my understanding and I think have been the most predictive are the ones that Vishal has come up with. He did a previous one that we're going to talk about called Matrix.
A (1:39)
Is it beyond the black box? But yeah, the Matrix beyond the black box.
D (1:44)
Actually we should put this in the notes for this. But the single best talk I've ever seen on trying to understand how LLMs work is one that Vishal did at MIT, which Hari Balakrishnan pointed me to, and I watched that. So. So he did that work and then he's doing more recent work that's actually trying to scope out not only how LLMs reason, but like it has some reflections on humans reason too. And so I just think he's doing some of the more profound work and trying to understand and come up with models, formal models for how LLMs reason.
