20VC Podcast Summary
Episode Title: Cohere's Chief Scientist on Why Scaling Laws Will Continue | Whether You Can Buy Success in AI with Talent Acquisitions | The Future of Synthetic Data & What It Means for Models | Why AI Coding is Akin to Image Generation in 2015
Host: Harry Stebbings
Guest: Joelle Pineau, Chief Scientist at Cohere
Date: November 3, 2025
Episode Overview
In this episode, Harry Stebbings interviews Joelle Pineau, the Chief Scientist at Cohere. Joelle, formerly VP of AI Research at Meta, is known for her pioneering work in reinforcement learning, robotics, and responsible AI development. The conversation spans the challenges of scaling in AI, the undervalued importance of algorithms versus data and compute, the role of synthetic data, talent acquisition, the evolving interface between humans and AI, and the societal implications of rapid AI progress and adoption—particularly in enterprise settings.
Key Discussion Points & Insights
1. Lessons from Meta & Research-to-Product Shift
- Time Horizons in AI Progress:
- Joelle reflects on her tenure at Meta, emphasizing how some breakthroughs in AI take years to mature despite the perception of “AI moving at the speed of lightning.”
“Sometimes how long it takes to prove out a hypothesis. We feel like AI is moving at the speed of lightning, but in fact there’s some things that it just takes a few years to mature...” (03:18)
- Joelle reflects on her tenure at Meta, emphasizing how some breakthroughs in AI take years to mature despite the perception of “AI moving at the speed of lightning.”
- Patience with Ideas:
- Despite hype cycles, many techniques (like reinforcement learning) have taken decades to reach usefulness in current agent/model paradigms.
“Sometimes you have to be a little patient with these ideas and the right algorithmic tweak, the right context, the right problem domain just opens up the magic.” (04:06)
- Despite hype cycles, many techniques (like reinforcement learning) have taken decades to reach usefulness in current agent/model paradigms.
2. Reinforcement Learning (RL): Challenges & Potential
- Fundamental Nature of RL:
- RL remains central conceptually but is not an immediate path to AGI due to inefficiencies and data requirements.
“Training through a system of rewards…is so fundamental. It’s not going away.” (04:47)
- RL remains central conceptually but is not an immediate path to AGI due to inefficiencies and data requirements.
- Why RL is Inefficient:
- Error compounding in sequential decision-making and the need for dynamic, expensive environments make RL costly.
“Every time you make a mistake, it sort of compounds… And to get it right is quite difficult. Sometimes people compare it to like a needle in a haystack.” (05:45)
- Error compounding in sequential decision-making and the need for dynamic, expensive environments make RL costly.
- Cost Evolution:
- Costs are decreasing where good reward functions exist (e.g., games), but for open-ended, social behaviors, progress remains hard.
3. Training vs. Inference Economics
- Market Allocation:
- Training has dominated AI spend, but inference (serving models) will eventually be a much larger share as enterprise models run on-premises.
“Enterprise…run it locally...motivation to have very efficient models so that they can run really efficiently on premise.” (08:37)
- Training has dominated AI spend, but inference (serving models) will eventually be a much larger share as enterprise models run on-premises.
- Unpredictability:
- The largest economic challenge is forecasting needs (data, compute, breakthrough timing), making AI a risky investment climate.
“It’s very hard to have predictability...Everyone wants to know how many GPUs do I actually need?” (10:07)
- The largest economic challenge is forecasting needs (data, compute, breakthrough timing), making AI a risky investment climate.
4. Algorithmic Innovation vs. Scaling Laws
- Linear vs. Nonlinear Progress:
- Compute and data generally give predictable, linear improvements; algorithms introduce nonlinear step changes.
“Algorithms are the ones that have the nonlinear effect...you can explore lots of ideas and then something like the transformer comes along and just changes the paradigm.” (11:12)
- Compute and data generally give predictable, linear improvements; algorithms introduce nonlinear step changes.
- Scaling Laws Endurance:
- Scaling laws still prove powerful, but rely on algorithms for step changes; data and compute alone aren’t everything.
“The scaling laws have been remarkably robust...but most of the time I wouldn’t bet against it.” (13:14)
- Scaling laws still prove powerful, but rely on algorithms for step changes; data and compute alone aren’t everything.
5. Research vs. Productization Tension
- Academic vs. Real-World Signal:
- Real enterprise adoption gives “tangible signal” on AI’s utility, beyond academic benchmarks.
“When you need to sell AI to a business, you get a real signal of what works, what doesn’t work.” (14:43)
- Real enterprise adoption gives “tangible signal” on AI’s utility, beyond academic benchmarks.
- Measuring Value in AI:
- Productivity gains (e.g., 10x more output per employee) are more realistic and significant than outright workforce replacement.
“Can most of your employees do 10x the amount of work with AI versus on their own? That to me is actually a better barometer.” (16:16)
- Productivity gains (e.g., 10x more output per employee) are more realistic and significant than outright workforce replacement.
6. Enterprise Adoption: Hurdles & Opportunities
- Integration Complexity:
- Integrating AI with legacy data/information systems is a massive challenge, especially given decades-old infrastructure.
“The challenge is to deploy in a way that allows them to exploit all of the information systems that they already have.” (20:42)
- Integrating AI with legacy data/information systems is a massive challenge, especially given decades-old infrastructure.
- Data Confidentiality & Security:
- On-prem deployment and data privacy are major concerns and opportunities in enterprise AI.
7. Security & Government Role
- AI Agent Vulnerabilities:
- Future agentic systems may face new threats such as impersonation (not just hallucination); need for robust standards/testing.
“Agents that come along and are essentially impersonating entities...taking actions on the behalf of these entities where they don’t legitimately represent...” (23:11)
- Future agentic systems may face new threats such as impersonation (not just hallucination); need for robust standards/testing.
- Regulatory Dynamics:
- Governments move slowly but can still set much-needed standards (analogy to aviation safety); companies must build at scale.
“Clear standards actually means reducing uncertainty for a lot of companies in this space.” (24:50)
- Governments move slowly but can still set much-needed standards (analogy to aviation safety); companies must build at scale.
8. Globalization & Team Building in AI
- Diversity Across Regions:
- The spread of AI R&D globally (e.g., Mistral in France, Cohere in Canada) encourages different approaches and serves local needs.
“Having a company that is attuned to that, that values that internationalization of model is actually important on the global market.” (27:51)
- The spread of AI R&D globally (e.g., Mistral in France, Cohere in Canada) encourages different approaches and serves local needs.
- Team Composition:
- You need a mix: a few visionary “Galacticos,” strong executors, and social glue. Money alone can’t buy great teams.
“You do need a few of these, like uber talents...but you don’t need all of your team.” (29:55)
- You need a mix: a few visionary “Galacticos,” strong executors, and social glue. Money alone can’t buy great teams.
9. Investments: Where to Spend and Why
- Balance Compute, Talent, Data:
- Data’s value (esp. specialized/synthetic) is growing relative to compute; building environments and benchmarks for training is increasingly costly.
“You need more specialized tasks...so that’s more expensive talent...also a lot of data that’s synthetic...generation of environments and benchmarks...can be pretty expensive too.” (32:14)
- Data’s value (esp. specialized/synthetic) is growing relative to compute; building environments and benchmarks for training is increasingly costly.
10. Synthetic Data and Model Degradation
- Risks and Exceptions:
- Synthetic data can lead to model degradation via loss of diversity in open domains, but is effective in bounded environments like games or, with care, in code.
“If you think like images, languages like LLMs talking to each other… you definitely get the degradation… due to a loss of diversity.” (35:41)
- Synthetic data can lead to model degradation via loss of diversity in open domains, but is effective in bounded environments like games or, with care, in code.
11. Code Generation: Where Are We Now?
- Analogy to Image Gen Progress:
- AI code generation is like image generation in 2015—low quality now, but likely to improve massively.
“We had image generation models in 2015. They were really bad...wait another 10 years...the quality of the code that’s produced is going to be excellent.” (37:31)
- AI code generation is like image generation in 2015—low quality now, but likely to improve massively.
- Curation Becomes Central:
- The new challenge will be picking “quality out of the volume”—the human’s role becomes intent, critique, and selection.
“That’s your 10x productivity improvement… But you still need people with intent.” (39:37)
- The new challenge will be picking “quality out of the volume”—the human’s role becomes intent, critique, and selection.
12. Interaction Paradigms: Prompts and Beyond
- Limits of Prompt Boxes:
- Current prompt-based interfaces are limited; future interfaces will be more multimodal—voice, gesture, gaze—but language remains core due to information density.
“The idea of like typing in a box… is very limited. And we're going to break out of that box.” (40:28)
- Current prompt-based interfaces are limited; future interfaces will be more multimodal—voice, gesture, gaze—but language remains core due to information density.
13. Shifting Beliefs & Bubble Dynamics
- Changing Convictions:
- Joelle once doubted neural nets would endure as the best paradigm but now sees their dominance due to backpropagation and gradient descent.
“Neural nets seem to be here to stay and...to do backpropagation and gradient descent...is a really powerful way to learn.” (41:31)
- Joelle once doubted neural nets would endure as the best paradigm but now sees their dominance due to backpropagation and gradient descent.
- Skepticism of Extremes in AI Risk:
- Criticizes both doomsaying and utopian AGI predictions as lacking rigor.
“I don’t have a lot of patience…for people…predicting sort of the extremist scenarios...” (42:49)
- Criticizes both doomsaying and utopian AGI predictions as lacking rigor.
- Is It a "Good" Bubble?
- High variance; rewards will be big but so will losses. Tolerance for risk is essential.
“As long as people have a tolerance to risk, then I think AI is a great investment…” (44:04)
- High variance; rewards will be big but so will losses. Tolerance for risk is essential.
14. Evaluation, Openness, and Education Access
- Evals as Indicators, Not Goals:
- Benchmarks are useful but shouldn’t be the only measure; enterprise cares about ROI, not Math Olympiad wins.
“Evaluation...gives you like a signal...But as we're building systems that are more and more general, do very specific tasks. You don't optimize for these.” (45:01)
- Benchmarks are useful but shouldn’t be the only measure; enterprise cares about ROI, not Math Olympiad wins.
- Education Disparity:
- Universities have less compute but still generate breakthrough ideas—ecosystem benefits from this diversity.
15. The Future: AI for Scientific Discovery & Efficiency
- Open Models Still Matter:
- Joelle warns against “closed world” isolationism; research thrives on openness.
“That’s a deep mistake. I will continue to believe that especially for research, the ideas need to circulate.” (54:51)
- Joelle warns against “closed world” isolationism; research thrives on openness.
- Excitement for Scientific AI:
- Looks forward to AI accelerating scientific discovery and efficiency, especially models runnable on modest hardware.
“AI for scientific discovery is going to be pretty fascinating to see…super keen to see what we’re going to be able to do at the scale that runs on like one or two GPUs.” (53:42)
- Looks forward to AI accelerating scientific discovery and efficiency, especially models runnable on modest hardware.
Notable Quotes & Memorable Moments
-
On Scaling Laws:
“The scaling laws have been remarkably robust. Lots of people have bet against scaling laws in the past and...I wouldn't bet against it.” – Joelle Pineau (13:14)
-
On RL Inefficiency:
“Every time you make a mistake, it sort of compounds...Sometimes people compare it to like a needle in a haystack.” – Joelle Pineau (05:45)
-
On Data's Growing Importance:
“You need more specialized tasks...that's more expensive talent...also a lot of data that's synthetic...generation of environments and benchmarks...can be pretty expensive too.” – Joelle Pineau (32:14)
-
On Prompt Interfaces:
“Typing in a box…is very limited. We're going to break out of that box...Language is incredibly powerful, but we're going to see much more multimodal ways to interact.” – Joelle Pineau (40:28)
-
On Extremist AI Narratives:
“I don’t have a lot of patience…for people…predicting sort of the extremist scenarios...I'm much more pragmatic, grounded. I'm pro innovation.” – Joelle Pineau (42:49)
-
On the Need for Openness:
“That’s a deep mistake...for research, the ideas need to circulate.” – Joelle Pineau (54:51)
Timestamps for Key Segments
| Timestamp | Topic | |-----------|---------------------| | 03:18 | Meta years: Patience in AI research | | 04:47 | Fundamentals and inefficiencies of RL | | 08:37 | Training vs. inference – enterprise focus | | 10:07 | Economic unpredictability in AI development | | 11:12 | Step changes: Algorithms vs. linear scaling | | 13:14 | Endurance of scaling laws | | 16:16 | Useful enterprise barometers for AI productivity | | 20:42 | Integration and challenges in enterprise AI deployment | | 23:11 | Security risks for agents and LLMs | | 24:50 | Government and standards in AI regulation | | 27:51 | Globalization and multilingual/sovereign model development | | 29:55 | The myth/reality of “buying” superstar teams | | 32:14 | Data’s growing expense and complexity | | 35:41 | Synthetic data’s risks to diversity and performance | | 37:31 | AI code vs. image generation analogy | | 40:28 | Future multimodal interfaces with AI | | 41:31 | Changing beliefs: neural nets’ dominance | | 42:49 | Skepticism of AI “existential risk” narratives | | 44:04 | The AI bubble: high risk, high opportunity | | 45:01 | Benchmarks: useful indicators, not ultimate metrics | | 53:42 | AI for scientific discovery and efficient models | | 54:51 | Open research in an era of closed models |
Tone & Style
The dialogue is candid, technical but accessible, with Joelle blending pragmatism and optimism. She continually emphasizes the need for patience, creative risk-taking, openness, and realism in both technological and societal discussions on AI.
For newcomers to the world of AI or veterans seeking up-to-the-minute thinking from top minds, this episode offers a grounded yet ambitious take on where the field is going—and why building for the long term, with an eye on both scaling and creativity, is the smart bet.
