In this episode of The Effortless Podcast, Amit Prakash and Dheeraj Pandey are joined by Alex Dimakis for a wide-ranging, systems-first discussion on the future of long-horizon AI agents that can operate over time, learn from feedback, adapt to users, and function reliably inside real-world environments. The conversation spans research and industry, unpacking why prompt engineering alone collapses at scale; how advisor models, reward-driven learning, and environment-based evaluation enable continual improvement without retraining frontier models; and why memory in AI systems is as much about forgetting as it is about recall. Drawing from distributed systems, reinforcement learning, and cognitive science, the trio explores how personalization, benchmarks, and context engineering are becoming the foundation of AI-native software. Alex, Dheeraj, and Amit also examine the evolution from SFT to RL to JEPA-style world models, the role of harnesses and benchmarks in measuring real progr...
Get AI-powered summaries and transcripts for any meeting, phone call, or podcast.
Available on iOS, Android, Mac, and Windows
Want AI summaries for your own recordings? Try Wave AI free →
No transcript available.