
Hosted by Matthew Harris · EN

Exuberance is coming back to venture, but the real action is happening in public markets first. This episode breaks down where AI stops being a story and starts showing up in earnings, margins, and cash flow. It traces how liquidity returns, why precautionary thinking keeps missing incentive driven adoption, and which companies are positioned to capture real value as AI moves from experimentation into core workflows. This is a forward looking map of the next risk on cycle, grounded in incentives, accounting, and competitive pressure, not hype. If AI really is a platform shift, this is where the scoreboard lights up.

2026 looks less like disruption and more like acceleration. In this episode, we map the upside, AI agents becoming premium tools, infrastructure spending unlocking new careers, Amazon and Google compounding quietly, and space opening as a real commercial frontier. This is about falling costs, rising capability, and a bigger economic surface area than we’ve had in decades.

**Are the state-of-the-art autoregressive decoder-style transformers the only future for large language models?** **We dive into the most fascinating alternatives, including linear attention hybrids that promise huge efficiency gains for long contexts and text diffusion models that generate tokens in parallel instead of sequentially.** Plus, discover how Code World Models are training models to simulate code behavior for improved modeling performance, aiming to develop more capable coding systems.

Claude 4.5, Sora 2 and more

In this episode of Agora, we explore how memory shortages, industrial policy, and national strategy are reshaping the future of compute. The global AI supply chain is under pressure, with high bandwidth memory emerging as the critical bottleneck that could slow the pace of model training. At the same time, SMIC is advancing production despite heavy restrictions, signaling that China is determined to push through technological ceilings. U.S. export controls add another layer of complexity, functioning as both a weapon of leverage and a risk of unintended acceleration for rivals. Together, these forces reveal a fragile equilibrium where geopolitics and hardware innovation collide.

AWS dominated cloud computing but watched the GenAI boom from the sidelines. Microsoft went all in on OpenAI. Google doubled down on DeepMind. Amazon looked behind. Until now.This episode covers the real story behind AWS’s AI resurgence, driven by its new powerhouse partner Anthropic. Billions in custom silicon orders. Massive multi-gigawatt training builds. A full-stack commitment to Trainium2, tuned for reinforcement learning at extreme scale.We unpack why memory bandwidth is the new battleground, how co-designed chips and models tilt the TCO equation, and why this partnership might redraw the map of AI infrastructure.This is not a pivot. This is a power move.

see more at Agora on Substack

In this episode, we dive into the high-stakes world of AI, starting with the impact of DeepSeek R1, a Chinese LLM that initially disrupted the market by undercutting leading models by over 90% on output token pricing. We'll explore the intriguing shift in user behavior, where DeepSeek's own web app and API service lost market share to third-party hosts, despite its low price. This shift illuminates the crucial role of tokenomics, revealing how a model's price per token is an output of key performance indicators like latency, interactivity, and context window, which DeepSeek actively trades off to save compute for its AGI research goals rather than user experience. Then, we pivot to Meta's aggressive pursuit of superintelligence, a strategy spurred directly by losing its open-weight model lead to DeepSeek. Discover how Mark Zuckerberg is personally driving this effort, reinventing Meta's datacenter strategy to prioritize speed with new "Tents" and building multi-gigawatt AI training clusters such as Prometheus and Hyperion. We'll also uncover the technical missteps behind the "epic fail" of Meta's Llama 4 model, including issues with chunked attention, expert choice routing, and data quality, and how Meta is addressing its talent gap by offering unprecedented compensation to top AI researchers.

This week on Agora, we dive into Ray Dalio’s proven framework for success, distilling the principles that propelled him to build Bridgewater Associates into a global powerhouse. Explore his “idea meritocracy” driven by radical transparency and unpack his five-step process for achieving goals: setting clear objectives, pinpointing problems, diagnosing root causes, crafting strategic plans, and executing with precision. We’ll also cover life and work principles that emphasize self-awareness, open-mindedness, and aligning your career with your core values. Tune in to gain practical insights for navigating reality and advancing your personal and professional growth.

• This excerpt introduces “Deep Dive: The Anthology of Balaji - A Guide to Technology, Truth, and Building the Future,” a book compiled by the same individual behind the “Almanac of Naval Ravikant.” • The text highlights Balaji Srinivasan as a prominent entrepreneur, investor, and futurist, suggesting that his ideas offer unique perspectives on identifying opportunities, breakthrough technologies, and constructing impactful ventures. • The book is presented as a comprehensive collection of Srinivasan’s most valuable and enduring ideas, gathered from various sources throughout his career. • The provided snippets offer insights into Srinivasan’s thoughts on the decline of state trust, the importance of independent thinking over conformity, a tiered understanding of leadership that champions technological advancement, and a staged process for developing ideas into successful businesses. • The source advocates for “building an alternative” as a response to perceived systemic flaws rather than merely critiquing them.