
Loading summary
A
Today on Upriers, we're talking with Liam Fetus. Liam is one of the co creators of ChatGPT, which I think almost everybody uses at this point. He was the VP of Post Training at OpenAI and before that was at Google Brain where he worked on a variety of really early AI innovations. Liam will be telling us a bit about Periodic Labs, his company which is focused on building an AI foundation lab for atoms. In other words, how do we impact the physical world, material sciences, chemistry, etc. Using AI. Very exciting topic and excited to be talking with them today. Leah, thank you so much for joining us today on no priors.
B
Yeah, thank you so much for having me. It's great to see you.
A
Yeah. So maybe what we can do. I think you're doing incredibly interesting things in terms of alternative types of models specifically for material sciences for the physical world. Effectively what you're building is an AI foundation lab for atoms, which I think is fascinating.
B
That's right.
A
But maybe we can start with is a little bit more of your background. You know, I think you were VP at OpenAI, you worked on one of the first trillion parameter models ever, et cetera. Could you tell us a little bit more about just like what got you here?
B
Yeah, so even further back, I was a physics major in undergrad, spent some time doing dark matter research. We had an apparatus that was directionally sensitive to dark matter's direction. So it was very interesting.
A
Why, why are they. Sorry, I'd love to come back to this, but why are there so many physicists in AI right now? So you look at Dario Modi, who runs Anthropic, of course.
B
Yeah.
A
You look at Adam Brown at Google, you look at a variety of people and they all kind of have these physics backgrounds.
B
Yeah. My old manager Joshua, also physics anthropic.
A
Yeah. Why do you think that is?
B
I think it's a great way to think about the world. It's like very principled, very like hard nosed scientists, very careful and I don't know, I think it's just, it's such an incredible field. You have such high leverage in computer science, in AI and so I think a lot of physicists were seeing that, particularly in like high energy physics. After the discovery of the Higgs. I think a lot of high energy physicists were sort of looking for what's next. Ultimately it becomes bottlenecked on the new apparatus for, you know, pushing the next energy frontier. And I think a lot of physicists were looking at their skillset and looking at the progress elsewhere and saying like Hey, I think I could be a huge contributor elsewhere.
A
It's just been fascinating to see like string theorists and people working on black holes and all sorts of effects, like kind of moving into AI.
B
Absolutely.
A
It almost feels like we're recreating the Manhattan Project or something. Except now what we're seeking is different forms of intelligence.
B
Yeah, that's right.
A
Kind of just from that perspective. Sorry to interrupt. So you know you studied physics, you worked on dark matter.
B
That's right. And then I was basically, and then in grad school in physics I was always gravitating towards the machine learning problems. I was looking at particle reconstruction and it's effectively machine learning problems. But it felt if I really wanted to push frontier of machine learning I should be in computer science. So ended up at Google Brain. Was overlapping with the first year residents there. Absolutely remarkable group of people. Remarkable period for Google Brain. I mean it's an era of when there's the creation of distributed training strategies, mixture of experts, the transformer. It was a really rich period in that history and it was a fun kind of like Cambrian era where people were really pushing the frontier with just like a handful of GPUs, really small collaborations. The field was much, much earlier and I think there was a lot of diversity and entropy in the research and it was very fun.
A
So it was kind of late 2010s or so, something like that.
B
This was 2016, 2017. So Google brand at that point was still really small and eventually was subsumed by DeepMind or combined with DeepMind. So I was at Google for many years mostly just doing architecture work. So was really pushing sparsity that allows for more efficient serving of models at scale and just really pushing the scale of what we could do. Towards late 2022 really became excited about the creation of products. The technology was getting very compelling and so I ended up that OpenAI with some other Googlers as well.
A
And what did you work on specifically at OpenAI?
B
Well, so the goal was we need to come up with some productionization of GPT4. So OpenAI had GPT4, it was pre trained and there were some Lerov post trains on it and there's questions about like okay, how do we turn this incredibly powerful model into products? And we're all spitballing ideas like writing bot, coding bot, you know, very natural. At the time some of our least interesting ideas were a meeting bot. So it would just sit in a Google Meet, take notes and then send out like to DOS after. But John Schulman was very opinionated he's like, we think we should keep it very general. Let's do chatbot. And that became a large part of the effort for those few months.
A
That's really cool. Yeah. So we worked on ChatGPT.
B
That's right.
A
And obviously I felt like that was kind of the starting gun of this whole AI revolution, or at least in terms of people's awareness. Like I'd started investing in the area beforehand.
B
Right. But.
A
But it seemed like almost as a secret up until ChatGPT came out. And then suddenly everybody realized that there's this powerful technology available.
B
Yes.
A
How did that lead you to materials and atoms and, you know, the physical world again? I know that was sort of your starting point in terms of economics, but what brought you back, given how much is being transformed right now through language?
B
I think just the inevitability of connecting these systems to the physical world. The opinion that I and others held periodic was you're not going to see the same kind of acceleration in science and technology unless you start connecting these things to the physical world. Science ultimately isn't sitting in a room thinking really hard. You have to conduct experiments, you have to learn from them, you have to interface with reality. And the creation of ChatGPT in late 2022 was a important technology, but it was still far too weak. We couldn't have done periodic on technology of that era. I think over the next few years past that, we saw ever improving models. We saw reasoning, I think like test time inference became really important. That led to more reliable error correction, more reliable tool use. And we see like the rise of coding agents and other agents. And I think those were foundational technologies necessary to then connect these systems to the physical world. It was just not possible with like the AI technology of 2022.
A
I guess the other thing that's missing from the physical world is data, or at least data that's easily accessible. So you look at something like the big foundation models on the language side and they're basically trained on the Internet as a major corpus. It's augmented in all sorts of ways with other data sources. How do you think about that for what you're doing, where you're trying to model atoms in the physical world and how all that stuff kind of works.
B
Yeah. So experiment, I mean, so we have simulation physics simulations and we have experiment and you know, I think exactly as you're pointing out, ML systems are good on the data you've trained them on, on the tasks you've trained them to do. I think sometimes there's like this mythology of AGI, asi, rsi, and I think we see increasingly powerful systems, but they do become limited if they don't have access to the raw data to actually make informed decisions.
A
How much data do you need? And so I know that there's some data scale related.
B
Yes.
A
Research and other things in terms of how you kind of hill climb towards like a really good model.
B
Yep.
A
How many experiments do you need to run or how many data points do you need or how do you think about the diversity of data points you need to generate? I'm a little bit curious, like what does that actually look like? Tangibly.
B
So there is some generalization from the existing models. So we don't need to reproduce a system that can understand and write English or write code. So we're kind of like leveraging.
A
And are you using open source for that or closed source models or some.
B
We use a combination. Yeah. So for example, like Periodic spends zero effort on improving coding models. We're incredibly impressed by Codex Claude code. And so that's been a huge accelerator for the company, but focused our machine learning efforts where the existing frontiers is not sufficiently good for us. I think going back to the data question, we're leveraging, call it order tens of trillions of tokens that went into open source models. And that's given this very foundational understanding. But once we start moving into specific discovery areas, chemical spaces, we can see a very high level of sample efficiency. So the system isn't starting as like a randomly initialized neural net. It has a strong prior on the world.
A
And so where does that prior come from? What data do you think that informs that?
B
Just general, just, just like, you know, papers, the Internet, as you're pointing out. Yeah, um, however, that's insufficient. Um, one of the engineers on our team was looking at a reported material property and it was just sort of extracted values from literature. And it was really interesting to see the reported value spanned many orders of magnitude. And so you train an ML system on that and it's like, well, the best you can do is model this distribution, but you're no closer to like a ground truth. And that's where experimental data comes in, where you now have a grounding in this. But really important, it's not just like a pool of data. It's this interactive closed loop system that is so powerful. Once you have the experimental data, you can look through it. You can look for aberrations, you can look for patterns, you can look for consistency with simulation data, with literature, and then that helps drive the next set of experiments. So it's not just a pool of data, it's a very active loop.
A
I see. And then how do you think about diversity data? So I look at something like Alpha 4 fold or some of the protein folding related models, which are amazing, right. If you think about it, I used to work as a biologist and we would, you know, a crystal structure would take years if it happened at all, because you wouldn't necessarily certain. If you could crystallize the specific protein under certain reaging conditions in a way that would be performant for actual, you know, crystal oxillography and everything, or NMR or whatever approach you took for structure. And then sort of alphafold comes out and you can just arbitrarily model anything right. On the protein world, which was, you know, amazing as a breakthrough, but it was a very specific data set that already existed, that had lots and lots and lots of structures. Over decades, over decades of work, how hard do you have to bootstrap that for every single materials domain? Or do you choose specific ones that you think can then generalize?
B
We have seen internally the greatest advances where we have an abundance of data in some space and that has led to the highest rate of acceleration internally. But I think you can think of different levels of generalization. And for systems that are strongly governed by quantum mechanical effects, there is some generalization there.
A
I see.
B
But if you produce a system that has modeled quantum mechanical objects really accurately, it's not really helping much on fluid dynamics or another level of abstraction. And so the generalization we're seeing is quite good. But there's almost a first principles you can.
A
Oh, that's so interesting. So you could do like here are the basic steps of chemical synthesis, quantum mechanics. Here's different aspects of how atoms interact in general or van der Waals forces or things like that.
B
Absolutely.
A
Oh, so interesting. Yeah, that's cool. And then from an architecture perspective, is there anything unique that you're doing or interesting or can you talk a little bit about how you're actually constructing some of these models on top?
B
Yeah. So language models are incredibly powerful. It's a very natural interface. And so we continue to use these, but we think about them almost as like an orchestration layer. So that's sort of a co pilot assistant, but also like a system that can direct experiments and it's orchestrating other specialized models as well. So we do construct neural nets that are specially designed for atomic systems where there's like some symmetry awareness and those have much lower latency and they've been like fine Tuned for that. And so basically you kind of think of this like, orchestrating layer that can ingest literature, it can go through our experimental data, it can go through different modalities, but they can also use specialized neural nets as tools, as reward functions. So it's like an overall system.
A
Okay, yeah, that makes a lot of sense.
B
Yeah.
A
I've seen a lot of people architect those sorts of approaches, even for things like customer support or other areas. Like, it seems like it's the common architecture that's emerging as you're doing these different use cases. That's right, yeah.
B
Yeah. But transformers have been very powerful.
A
Yeah, yeah. And that's really cool. So if I look at the language world, one of the things that was pretty unique about it, and it's the reason that I think these companies like OpenAI, Anthropic and others are growing so fast, is it just plugged into a very big domain of human existence, which is all language. And all language means enterprise software and enterprise interactions, and it means consumer behavior is basically how we interact with the world. Yes, it seems like there's a little bit more of a leap for other areas. So, for example, in robotics, there's really interesting things, different types of robots that exist in the world, but the footprint of that is quite limited relative to language. And the same seems to be true for material sciences. So how do you think about where you're going to commercialize this first, or who you're going to work with? Or are there specific domains of products that you're working on first?
B
So we've begun working very closely with scientists. We've treated periodic as our customer zero and seeing how can we transform how this field of science is done. But there's huge opportunities across all of these industries, all these enterprises that are interfacing with the physical world, people who are bottlenecked by materials engineering, process engineering, and again, those are kind of this, like the same natural interfaces where engineers are asking questions about their data, they're trying to find aberrations, they're trying to debug machinery, they're trying to get to a better formulation. It's actually a quite universal thing as well. And so we've kind of created our little testing ground internally, and now we're sufficiently excited about the tech we've been building and to see this acceleration for advanced manufacturing more broadly.
A
And is your model going to be developing materials for other third parties? Is it developing your own materials or that you then sell on the market, like, because it almost reminds me a little bit of a biotech model.
B
Yeah.
A
Where in biotech you can either partner with a big pharma and then effectively help them create a drug and take a royalty on it, or you can build your own drugs. How do you think about that in the context of what you're doing?
B
We're thinking about ourselves as an intelligence layer for these companies. So you can think about system of record, control plane for different experiments and getting to solutions. But like you're saying, there is a very interesting aspect of some breakthroughs here, could have really high value and it might be more akin to a discovery model like we've seen in biotech and elsewhere, but starting thinking about our just as a software business.
A
Have you ever had the diamond age?
B
Very fast.
A
Yeah. Have you had the diamond age?
B
No, I haven't.
A
It's the Neal Stephenson book. It's basically this book about. It was written in the 90s and there's two key concepts in it. One key concept is there's effectively an AI tutor that's unleashed on the world. And it kind of teaches huge numbers of young girls all sorts of skills. And it's this very interesting thing about AI education. And then in parallel, why in particular, basically this AI research scientist creates a primer for his daughter and the Chinese steal it and clone it and distribute it across the country. And because he built it for young girls, it suddenly every young girl in China has it.
B
Right, right.
A
So that's the reason it's very China theft of IP kind of thing. And then the other part of the book is, is about matter pipes into everybody's homes and they all have 3D printers and you download blueprints and it just creates whatever you need in the physical world. And some people start evolving different nanobots to do different things. It's this very advanced kind of AI plus materials kind of future world.
B
Yes.
A
What is your vision or conception of what our world looks like in 10 years, assuming periodic is successful?
B
Well, I mean, I think, as you're pointing out, you're going from systems that aren't just writing essays, not just writing software, but to literally generating matter. And I think it has pretty profound implications to semiconductors, aerospace energy. And I think it's incredibly important for can we increase the pace of just the physical development of the world? I mean, we see how quickly the digital realm is changing. Software engineering now looks wildly different than even six months ago, but I think we see similar opportunities in the physical world. Of course, atoms are hard, and so you will have some limits of physics, but just because atoms are hard doesn't mean there's not an order of magnitude or 2 to speed up just making sense of huge amounts of data and getting to solutions more quickly. Yeah. So I think what we're trying to do is give humanity this agency for atomic rearrangement synthesis, and we think it's going to just be a huge accelerator. So I mean, if our physical world could keep up at some fraction to our digital world, I think life will just feel dramatically different.
A
Yeah, it's kind of the revolution that could really come. Yeah, it kind of reminds me of almost the materials equivalent of the agricultural revolution. Yeah, we suddenly had a massive spike in productivity of alpha.
B
Exactly.
A
And it seems like there's been all sorts of bottlenecks that have constrained us until now that you fix what we're trying to address. That's right, yeah. What aspect of the work that you're doing are you most excited about the
B
iteration between these groups of people? I mean, this is just irreducibly a multidisciplinary problem. We have physicists and chemists working really closely with some of the top AI researchers in the world, working closely with some of the best engineers in the world. And this multidisciplinary close collaboration is just absolutely incredible because seeing firsthand how a field can fundamentally change people who have been doing research for in some cases decades in a field and now seeing like, oh, under these systems, under intelligent systems, it could look this very different way. And I mean, I use like an analog to machine learning a lot, going back to the early Google brain days where the frontier is pushed forward by a few GPUs and a few people. Now you look at this era where it's really like industrialized and there's dozens, hundreds of researchers working together with hundreds of Thousands, millions of GPUs dictated and driven by scaling laws. Everything is about scaling. It's given that predictability, it's allowed us to put huge amounts of capital into this field. And I think the physical sciences, physical engineering will have a very similar property where we establish these scaling properties and bring that mindset and so periodic in this field is really thinking about how do we bring much larger scale sets of experiments to bear on this. And intelligent systems have enabled us, automation has enabled us. And you really need both an improvement to automation where you can soon become create bottlenecks in intelligence. And I mean the scientists very much feel this where they're not used to working at that level of throughput and they just can't simply make sense of so much data.
A
So. Interesting. Yeah. So I Guess in terms of scale here, one of the real benefit, one of the things that's really benefited the, the, the frontier labs on the LLM side is just scale of capital and therefore scale of GPU and scale of data, of course. Is this similarly a capital intensive area in your mind?
B
Yeah, we will require more capital. GPUs are so extraordinarily expensive. And what's interesting is just the compute costs relative to physical infrastructure is actually surprising where so much money is spent on the compute that the physical infrastructure sometimes is actually lower but has very large lead times. And there's intrinsic difficulty of having these well calibrated, well functioning physical systems. Um, but from a capital perspective, it's primarily a compute cost.
A
Yeah, it's really interesting if you look up the cost of a Stanford postdoc, for example, relative to a machine learning engineer, it's like such a big difference. And you, you really, you know, my takeaway is that many people working in science, particularly in academic setting, are very undercompensated relative to sort of their societal value.
B
Absolutely.
A
And so I always like it when companies kind of help bring people into the, into the fold in terms of both human impact, but also that ability to do things at real scale and really do things a different way. So it must be very exciting for the people on your team.
B
Yeah. Some of the scientists who've joined us are among the best in the world and it's been absolutely incredible working with them.
A
Yeah, I mean it sounds like you've built such an amazing interdisciplinary team. Are there specific roles that you're actively looking for right now or key things that you really want to hire up?
B
Absolutely. So on our site we have decomposed the world into bits and atoms. It's a loose taxonomy. But on bit side we're really thinking about mid training, pre training roles from the AI side, always more infrastructure roles and on atom side like control engineering, system engineering, but also now thinking too about spanning that with product engineering across the.
A
So I think one of the things that everybody's really thinking deeply about or is excited about right now is AGI, asi, sort of these advanced systems that are as good as humans or better than humans at different things, or are very generalizable in terms of their abilities to do a broad swath of things. How do you think about that within the context of what's happening over the overall foundation model curve? Because obviously you were very integral in terms of the development of some of these systems. And then how do you think about that applied specifically to some of the areas you're working in.
B
I think one fallacy is thinking about intelligence as a scalar. We've consistently seen these systems have a very odd spikiness and it's actually possible to architect a system that is world class on some math domain, but then you could do some perturbations to the questions and actually degrade it substantially. So it's like a bad high school student. And so there's this odd spikiness to these systems.
A
So basically you can make a system that's like a genius at one thing and not very good at a bunch of other stuff.
B
And I guess the point I was making is those fields can actually be quite adjacent. So sometimes the generalization can be non intuitive. But one way I think about recursive self improvement is really kind of akin to neural architecture search from roughly 10 years ago. And I think there is a very clear path for software engineering. So these systems have become so incredibly impressive on this, on this domain as a result of huge amounts of data, really cheap verifiable environments like you know, you can check unit tests, go from failing to passing with just a few CPUs. It's basically instantaneous. There's no domain expertise gap between an AI researcher, software engineer and obviously this will become and is becoming a larger contributor to the next generation of the system.
A
When do you think it just flips into we just everything is machine self improvement versus human directed or needs a lot of human intervention. So do you think that's two years away? Do you think that's five years away? Do you think it's ten years away?
B
Well, I guess like building on what I was saying is I think there's a domain caveat to that.
A
Sure.
B
So rolling forward that software engineering self improvement, I think you're going to have a system that can write complete repositories, identify bugs, refactor code, but it doesn't suddenly understand biology.
A
Sure.
B
Right. It's just like there's a domain gap there in knowledge.
A
Yeah.
B
But even beyond that there's sets of strategies done in software engineering that differ from scientific or engineering strategies. So it's, you're not operating under, it's not like decision making under uncertainty to the same degree. It's like very verifiable and that's driven so much of our work. Um, so in that domain I think it's happening now. Ish. And I think we'll see the same thing too for AI research that's a slower outer loop because now the experiment isn't just checking some unit Tests passing, but it's checking. What was the scaling property? Did this model converge? What's the generalization of the system that requires GPUs, that requires, you know, many hours of experiments. But I think that will also will.
A
And those are all evals that people use today as they're looking at existing models. And so they do have that utility function, that feedback loop that can be just driven by self learning.
B
That's right, that's right. But again, like the connection of these things to the physical world is going to be so critical because both of those systems are being trained in a closed loop against that domain. So it's a closed loop for doing software engineering, a closed loop for doing AI research. And that's the premise of Periodic. Like we need to have these closed loops of action actually doing science, of actually doing engineering. And these two, I mean these two domains are how I think the rest of the world will go with some delay. And this is again like the foundational technology that we're super interesting.
A
Do you think you need sufficiently good robotic systems in order to have that closed loop for what you're doing? In other words, do you need something like PI or skills or something else to work in order for Periodic to hit that escape velocity? In terms of a closed loop system,
B
no, but it's a huge accelerator. The goal for Periodic is to generate high quantity, high quality data, diverse data, and automation is assistance to that. So right now we employ people as well and we have autonomous parts that are just, you know, very reliable. If you had a dexterous humanoid who could wander into an unstructured lab and make sense and follow instructions reliably, that would be a huge accelerator. Right now the automation of physical systems is, requires a very careful design and it's slow. But I think with improvements in robotics, this is going to accelerate this. But already the reliability of these sort of like hybrid systems is sufficient to produce huge amounts of reliable data. But it's just going to accelerate us further.
A
Yeah. One of the reasons I ask is I used to run this company, Color, and we built our own liquid handling robotic systems. Right. We'd buy liquid handling robots, but then we'd have to adjust them dramatically. We had like cameras that would use ML to monitor the system and sort of make adjustments. We had to 3D print parts to decrease vibrations on the platform because we were dealing with such small volumes of liquid.
B
Right.
A
And so there's enormous amounts of customization versus just having. And the firmware for it was awful. And writing against that was painful versus just Having like a robotic system that would work like a modern system in all the ways that you'd conceive that.
B
Right.
A
And so that's what the reason that I was asking is. If you really want to do high throughput experiments, you need these underlying systems to be able to do all the liquid handling and to do all the titration of stuff and all the rest of it, so.
B
Yeah, that's right. I mean, I think it's, look, right now we're using almost like more like off the shelf robotics. It's very simple, very commoditized, not doing a huge amount of innovation on that front. But again, as these more general robotic systems come to hit this reliability threshold, it's going to be a massive accelerator for spinning up new labs as well.
A
Yeah, you've seen such a wide range of different things happen in the AI world since indeed your work at Google, I guess, at this point about a decade ago. And so you were there during the birth of the transformer model, you were there for the birth of ChatGPT. What are you most excited about outside of periodic over the next few years in terms of what's happening with AI?
B
I mean, of course, robotics. Again, I'm just so excited about the interface of AI systems with the physical world. And we're approaching one angle of that, which is science, engineering, and we need that data in order to make those advances. But simply just agency and control of the physical world via robotics is going to be transformative. So I'm very excited about these interface layers. I think that's going to be such a massive opportunity because I mean, how many software engineers are there in the world versus people who like the physical world and there's just labor shortages everywhere. So yeah, I think it's going to be a very interesting decade.
A
Amazing. Well, thank you so much for joining us today.
B
Yeah, thanks so much. It was really great chatting today.
A
Yeah.
C
Find us on Twitter at nopriorspod. Subscribe to our YouTube channel. If you want to see our faces, follow the show on Apple Podcasts, Spotify or wherever you listen. That way you get a new episode every week and sign up for emails or find transcripts for every episode@no-priors.com.
Release Date: April 3, 2026
Hosts: Elad Gil & Sarah Guo
Guest: Liam Fedus, Co-Founder of Periodic Labs, former VP of Post Training at OpenAI
This episode explores how artificial intelligence is impacting materials science and the physical world through the work of Periodic Labs, co-founded by Liam Fedus. Drawing from his background in both physics and key roles at Google Brain and OpenAI, Fedus and the hosts discuss the intersection of advanced AI with experimental science, challenges unique to material engineering, and the implications if the “AI for atoms” revolution succeeds.
[00:52–03:33]
[03:33–05:14]
[05:14–06:34]
[06:34–09:52]
[09:52–11:29]
[11:29–12:34]
[12:45–14:41]
[15:10–17:48]
[17:48–21:16]
[21:44–24:55]
[25:39–27:59]
[27:59–29:01]
“It's such an incredible field... I think a lot of physicists were seeing that... I could be a huge contributor elsewhere.”
— Liam Fedus [01:38]
“It almost feels like we're recreating the Manhattan Project... except now what we're seeking is different forms of intelligence.”
— Host [02:29]
“The opinion... at Periodic was you're not going to see the same kind of acceleration in science and technology unless you start connecting these things to the physical world.”
— Liam Fedus [05:26]
“It was really interesting to see the reported value spanned many orders of magnitude. And so you train an ML system on that and it's like, well, the best you can do is model this distribution, but you're no closer to a ground truth.”
— Liam Fedus [08:52]
“If our physical world could keep up at some fraction to our digital world, I think life will just feel dramatically different.”
— Liam Fedus [16:16]
“One fallacy is thinking about intelligence as a scalar. We've consistently seen these systems have a very odd spikiness...”
— Liam Fedus [22:10]
“Agency and control of the physical world via robotics is going to be transformative.”
— Liam Fedus [28:19]
Summary prepared for listeners seeking a comprehensive understanding of the episode’s themes, breakthroughs, and candid insights.