Podcast Summary: Latent Space: The AI Engineer Podcast
Episode: The AI Coding Factory
Date: May 29, 2025
Guests: Matan and Eno, Co-founders of Factory AI
Hosts: Alessio (CTO, Decibel) and Wix (Founder, Small AI)
Episode Overview
This episode features a deep dive into Factory AI, a company redefining software development with AI-driven autonomous agents ("droids") for enterprises. The founders, Matan and Eno, join hosts Alessio and Wix to discuss the unique journey of Factory’s founding, the evolving landscape of AI agents and code generation, product positioning, enterprise use cases, technical architecture, and their vision for the future of AI-assisted software engineering.
Key Topics and Discussion Points
The Founding Story of Factory AI
Timestamps: 00:16 – 06:33
- LangChain Hackathon Origins: Matan and Eno met at a 2023 LangChain hackathon, bonding over a shared fascination for code generation and AI in software. Both had Princeton connections but had never meaningfully interacted before.
- “We had like 150 mutual friends, but somehow never had a one-on-one conversation... at this LangChain hackathon...very quickly just gets into code generation.”
– Matan [00:35]
- “We had like 150 mutual friends, but somehow never had a one-on-one conversation... at this LangChain hackathon...very quickly just gets into code generation.”
- Career Transitions: Eno was at Hugging Face focused on code model advising; Matan was doing a PhD in theoretical physics at Berkeley but was "nerd sniped" into AI by the fundamental nature of code as a lens for machine intelligence.
- “The beauty of how code is just core to the way that machines, you know, develop intelligence, really kind of nerd sniped me and got me to leave what I was pursuing for 10 years.”
– Matan [02:30]
- “The beauty of how code is just core to the way that machines, you know, develop intelligence, really kind of nerd sniped me and got me to leave what I was pursuing for 10 years.”
- Rapid Commitment: Within eight days of meeting, both founders quit their previous roles to start Factory AI, driven by an “intellectual love at first sight.”
- “It was eight days from us first meeting to me dropping out of my PhD and Eno quitting his job.”
– Matan [06:18]
- “It was eight days from us first meeting to me dropping out of my PhD and Eno quitting his job.”
Evolving the AI Coding Landscape
Timestamps: 06:56 – 14:25
- Market Positioning: Factory’s focus is on autonomous, end-to-end systems for software development in the enterprise, targeting legacy, large, messy codebases (e.g., COBOL migrations) often ignored by other AI coding tools.
- “There are hundreds of thousands of developers who work on code bases that are 30+ years old...the value you can provide is very dramatic.”
– Matan [07:14]
- “There are hundreds of thousands of developers who work on code bases that are 30+ years old...the value you can provide is very dramatic.”
- Constraints of Existing Coding Tools: Most competitors center on IDE plugins for individual productivity, facing constraints in latency and context. Factory’s product is cloud-first, enabling more scalable, delegative workflows.
- “When you are freed of a lot of these constraints, you can start to more fundamentally reimagine what a platform needs to look like...the product experience of delegation is really, really immature right now.”
– Eno [08:32]
- “When you are freed of a lot of these constraints, you can start to more fundamentally reimagine what a platform needs to look like...the product experience of delegation is really, really immature right now.”
- “Droids” Not Agents: Factory’s agents, dubbed “droids,” are designed around planning and environmental grounding rather than endless, unreliable loops. The naming avoids the baggage of the “AI agent” hype.
- “Droids have a nice ring to it...our customers really love droids as a name—‘these are the droids you’re looking for’.”
– Matan [13:16]
- “Droids have a nice ring to it...our customers really love droids as a name—‘these are the droids you’re looking for’.”
- Division of Labor: Founders see the core of software developer work as shifting; AI will take more of the “inner loop” (code writing), with humans increasingly focused on planning and communication.
- “The outer loop of software development… is going to continue to be very human driven, while the inner loop… is probably going to get fully delegated to agents very soon.”
– Eno [13:51]
- “The outer loop of software development… is going to continue to be very human driven, while the inner loop… is probably going to get fully delegated to agents very soon.”
Product Demo and Technical Differentiators
Timestamps: 14:37 – 36:19
-
Droid Specializations: Factory’s platform offers different autonomous agents (droids) tailored to key enterprise use cases:
- Knowledge Droid: Research, documentation, and technical writing.
- Code Droid: The “daily driver” for code changes, ticket execution, with workflow delegation.
- Reliability Droid: Popular for incident response and SRE work.
- “There are different droids available for key use cases that people tend to have...we have a Code Droid, Knowledge Droid, and a Reliability Droid.”
– Eno [14:37]
-
Workflow Design: Emphasis on seeing agent thought processes ("X-ray into its brain") and integrating context from Slack, Linear, Jira, GitHub, Sentry, PagerDuty, etc.
- “As the agent is working, what matters most is seeing what the agent is doing and having a bit of like an X-ray into its brain.”
– Eno [16:00]
- “As the agent is working, what matters most is seeing what the agent is doing and having a bit of like an X-ray into its brain.”
-
Intelligent Delegation: Agents ask clarifying questions dynamically, shifting from prompt engineering dependence toward more natural, manager-like delegation.
- “A lot of users...should not need to prompt engineer agents...the system knows when to ask for clarification.”
– Eno [18:23]
- “A lot of users...should not need to prompt engineer agents...the system knows when to ask for clarification.”
-
Proactive Context Synthesis: Factory’s platform generates synthetic insights into codebases (e.g., setup steps, module interconnections), reducing friction and preventing rote context ingestion.
- “As we index code bases, we’re actually generating these insights at a much more granular level across the entire code base. Systems should be proactive in finding that information.”
– Eno [21:36]
- “As we index code bases, we’re actually generating these insights at a much more granular level across the entire code base. Systems should be proactive in finding that information.”
Model Evaluation, Pricing, and Enterprise Integration
Timestamps: 24:10 – 44:14
- Model Benchmarks: Internal evaluation suite built on task-based and behavioral specs, rather than just public benchmarks ("Big Bar vs Little Bar" syndrome).
- “There are so many customers that we have. That was purely because they saw the charts...so I think that motivates resources to be put on benchmarking.”
– Matan [25:42] - “Vibe-based...internally actually matters a lot. We use factory every day, so when we switch a model we very quickly get a sense of how things are changing.”
– Matan [26:21]
- “There are so many customers that we have. That was purely because they saw the charts...so I think that motivates resources to be put on benchmarking.”
- Handling Model RL Preferences: Some new LLMs (e.g., Sonnet 3.7) appear to prefer specific coding styles or tools due to their RL post-training. Factory adapts its product to ensure consistency and optimal tool usage.
- “It smells like Claude code...what if you gave it a search tool that was way better than grep, but the model just loves to use grep?”
– Eno [27:00]
- “It smells like Claude code...what if you gave it a search tool that was way better than grep, but the model just loves to use grep?”
- Pricing Model: Usage-based, transparent billing on tokens consumed.
- “We’re fully usage based...I actually think that we get better users the more they understand what tokens are and how they’re used, you know, in each back and forth.”
– Matan [36:23]
- “We’re fully usage based...I actually think that we get better users the more they understand what tokens are and how they’re used, you know, in each back and forth.”
- Enterprise Metrics: Focused on deliverable timelines and concrete ROI over traditional productivity metrics like code churn or number of commits.
- “At the end of the day, no one really cares about the metrics. What people really care about is developer sentiment...pulling in timelines is the best ROI.”
– Matan [39:53]
- “At the end of the day, no one really cares about the metrics. What people really care about is developer sentiment...pulling in timelines is the best ROI.”
The Changing Nature of Software Development and Team Structures
Timestamps: 31:07 – 44:14
- Why Browser-Based (Not IDE): Factory intentionally eschews IDE plugins for a web-first interface, betting that as AI does more code generation, the optimal developer experience will change fundamentally.
- “Can you iterate your way from a horse to a car?...you do need to think from scratch, about what does that new way to develop look like.”
– Matan [31:07]
- “Can you iterate your way from a horse to a car?...you do need to think from scratch, about what does that new way to develop look like.”
- Rise of Tiny Teams and The “AI Native” Attitude: Individual users and very small teams can now accomplish work previously requiring dozens or hundreds of engineers.
- “There are sometimes individuals who weren't even really developers who will use Factory and have more usage than in 100 person enterprise...crazy to see.”
– Matan [58:34]
- “There are sometimes individuals who weren't even really developers who will use Factory and have more usage than in 100 person enterprise...crazy to see.”
Bottlenecks, Future Vision, and Organizational Insights
Timestamps: 47:51 – 57:42
-
Technical Limiters:
- For Models: Need for LLMs capable of long, complex, goal-oriented agentic tasks (multi-hour sessions with persistent planning).
- “Models that have been post-trained on more general agentic trajectories over very long time spans...that is probably one of the bigger blockers.”
– Eno [48:27]
- “Models that have been post-trained on more general agentic trajectories over very long time spans...that is probably one of the bigger blockers.”
- For Dev Tools: More robust, semantic observability and analytics are still lacking; current tools only offer rudimentary traces.
- “It is still surprising to me that observability, it remains very challenging... how do you build almost semantic observability into your product?”
– Eno [50:03]
- “It is still surprising to me that observability, it remains very challenging... how do you build almost semantic observability into your product?”
- For Models: Need for LLMs capable of long, complex, goal-oriented agentic tasks (multi-hour sessions with persistent planning).
-
Customer Growth and Go-To-Market: Factory’s enterprise deployments are growing rapidly on the strength of "aha" moments with customers; go-to-market investments are ramping up.
- “We really just relied on word of mouth...and when every one of those ends up a happy customer, you need to increase top of funnel.”
– Matan [52:24]
- “We really just relied on word of mouth...and when every one of those ends up a happy customer, you need to increase top of funnel.”
-
Unique Hiring Needs: Biggest challenge is finding highly technical people who can both interface with execs and dig into code hands-on ("Is this a junior Eno or not?").
- “I think a big rate limiter is...having both that ability to talk to the CIO, VP Engineering... and sit side by side with their developers and jump into the platform.”
– Matan [53:52]
- “I think a big rate limiter is...having both that ability to talk to the CIO, VP Engineering... and sit side by side with their developers and jump into the platform.”
-
Importance of Brand, Vibe, and Team: Factory benefited from close collaboration with a design-minded team (including Matan’s brother), stressing the importance of cross-disciplinary creativity and creating a company culture that’s fun and social as well as ambitious.
- “I recommend working with a sibling...having that design perspective and the engineering perspective and bash those two things together until we get something perfect.”
– Matan [55:19]
- “I recommend working with a sibling...having that design perspective and the engineering perspective and bash those two things together until we get something perfect.”
Notable Quotes & Memorable Moments
- “It was eight days from us first meeting to me dropping out of my PhD and Eno quitting his job.”
– Matan [06:18] - “There are hundreds of thousands of developers who work on code bases that are 30+ years old...if you made a demo video doing some COBOL migration, that's not very sexy...but the value you can provide is very dramatic.”
– Matan [07:14] - “The product experience of delegation is really, really immature right now. And most enterprises see that as the holy grail, not just going 15% or 20% faster.”
– Eno [08:32] - “A lot of users, we believe, should not need to prompt engineer agents...if you're hyper optimizing every line...you're going to have a bad time.”
– Eno [18:23] - “We are taking that more ambitious angle...everything is going to change about software development...the time developers spend writing code is going to go way down. But the time spent planning is going to go way up.”
– Matan [31:07] - “No one really cares about the metrics. What people really care about is developer sentiment...pulling in timelines for big deliverables.”
– Matan [39:53] - “If you are highly, highly technical but you want to be a founder...interface with CIOs and CTOs, this is a huge opportunity.”
– Eno [54:26] - “I cannot recommend enough working with a sibling... having that design perspective and the engineering perspective and bash those two things together.”
– Matan [55:19]
Important Timestamps
| Segment | Timestamp | |-----------------------------------------------|----------------| | Founding story/hackathon meeting | 00:16 – 06:33 | | Autonomous agent focus & "droids" concept | 06:56 – 14:25 | | Product demo, use cases, and workflow | 14:37 – 36:19 | | Model evaluation and pricing | 24:10 – 44:14 | | Browser vs. IDE paradigm shift discussion | 31:07 – 36:19 | | Metrics, ROI, and enterprise deployment | 36:19 – 44:14 | | Bottlenecks and future model vision | 47:51 – 50:03 | | Organizational growth, hiring, and design | 51:56 – 57:42 | | Closing comments: AI-native teams, culture | 57:42 – 59:09 |
Takeaways & Key Insights
- Factory AI is betting big on the shift from collaborative, IDE-centric code writing to a cloud-based, delegative model—the "AI coding factory."
- Their platform is built for enterprises and real-world, often unsightly codebases—eschewing hype demos for high-impact, unsexy tasks.
- The team is obsessive about product experience (especially explainability, workflow, and delegation) and sees a radical change ahead for the developer role.
- Growth is accelerating among Fortune 500s; word-of-mouth and demonstration of dramatic ROI (shortening massive migrations from months to days) are key.
- Technical and go-to-market scaling both hinge on hybrid technical-sales hires; brand and team culture are also seen as strategic superpowers.
For full show notes and resources, visit latent.space.
