Transcript
A (0:00)
Today on the AI Daily Brief, everything we know so far about GPT5 before that in the Headlines Vibe Coding is finally coming to the Enterprise. The AI Daily Brief is a daily podcast and video about the most important news and discussions in AI. All right friends, quick announcements before we dive in. 1. Thank you to today's sponsors, Blitzy, Plum and Vanta. To get an ad free version of the show go to patreon.com aidaily Briefly. Second, if you are interested in sponsoring the show, we are starting to fill up for our fall and winter slots. Shoot me a note at nlwbreakdown.network and I can send you more information. Lastly, I mentioned this yesterday, but if you are a dev shop or a build agency that actually builds agents, I want to hear from you. Superintelligent is of course routing people to the use cases that are most relevant for their companies. But then they gotta build things. And if you have the capabilities to build things, especially at meaningful enterprise scale, shoot me a Note@NLWSuper AI with agent builder in the subject line because I want to hear from you. But with that, let's get into today's episode. Welcome back to the AI Daily Brief Headlines edition. All the daily AI news you need in around five minutes. We kick off today with something that was on the one hand completely inevitable, but on the other hand is still extremely exciting, which is the fact that Vibe Coding is finally coming to the enterprise in a major way. Replit and Microsoft have announced a strategic partnership that will see replit added to Microsoft's Enterprise Cloud Store. The platform will also be integrated into Microsoft's cloud services like Containers, virtual machines and their version of postgres. This integration will allow app builders to have easy access to enterprise grade backend infrastructure while allowing Microsoft to earn revenue from the developer ecosystem. Now, while Microsoft already offers their own AI coding platform in the form of GitHub Copilot, they say that Replit will be a complimentary addition to their product lineup. Instead of a replacement for Copilot, they're pitching RELET as a substitute for no code prototyping and design tools like Figma. The pitch is that non technical business managers can use replit to build their own apps. Deb Kupp, the president of Microsoft America, said, at Microsoft, we believe every person in every organization should be empowered to achieve more through technology. Our collaboration with ReKit democratizes application development, enabling business teams across enterprises to innovate and solve problems without traditional technical barriers. Replit CEO Amjad Massad added, we aspire for replit to be the most trusted name for enterprise in this new era of agentic coding. Now TechCrunch noted that there is one big loser from the new partnership, writing, if there is any competitor taking an L from this partnership, it's Google Cloud. The apps built and run through replit are typically hosted on Google Cloud. In fact, replit has been such a feather in Google's cap that the cloud giant has profiled the partnership. However, the deal is non exclusive, meaning that the startup is not leaving Google Cloud but is growing to support Microsoft Shops. Still, to me, from where I'm sitting, this is a completely obvious type of tie up and I agree wholeheartedly with Karthik Harryharan who writes I predict Microsoft will acquire a company in the AI vibe coding space within the next six to 12 months, just like their acquisition of GitHub in 2018. Vercel, Lovable, Replit, et cetera will all be acquisition targets. To me this is just incredibly obvious. We are already seeing certainly in the startup domain, but also creeping into the enterprise how these vibe coding tools are changing how people do their work, especially in non technical roles. But enterprise vibe coding really is a different, more complicated beast than consumer level vibe coding. It's going to take someone like a Microsoft who has deep pockets and patience to create a context where it's worth pursuing. But boy is the pot of gold at the end of that rainbow a big one. Next up today, some funding news. Mistral is in talks to raise a billion dollar funding round as they grow into a regional champion in Europe. The French AI company is reportedly in talks to raise a billion dollars in equity from several investors, including Abu Dhabi fund mgx alongside, the firm is seeking hundreds of millions of euros in debt capital from French lenders, including BPI France. To date, Mistral has raised around a billion euros since its 2023 founding, reaching a valuation of just under 6 billion after around last year. This round then would be a significant jump in available capital, enabling the firm to pursue European data center projects and compete in the next generation of foundation models. The fundraising also has a distinctly geopolitical angle. French President Emmanuel Macron has said that Mistral is central to the idea of European AI sovereignty. This deal would deepen ties between France and the uae, ensuring some level of independence from China and the US in AI competition. MGX already partnered with Mistral and Nvidia to construct Europe's largest data center campus in May. The 8.9 billion euro facility will be built in France and is projected to be operational by 2028 beyond that, the UAE has also committed to spend 50 billion euros on AI projects in France in support of Macron's push for AI sovereignty. In other funding news agent tooling platform LangChain is about to become a unicorn. According to TechCrunch sources, the company is set to raise a new round of funding at a billion dollar valuation led by IVP. LangChain started life in late 2022 as an open source project self funded by founder Harrison chase. In early 2023, as developer interest grew, Chase transformed the project into a startup. He closed a $10 million seed round from Benchmark that April and a week later he raised another 25 million in a Series A led by Sequoia, which valued the startup at 200 million. Lang chain was a breakout hit early during this AI era. It provided agentic tooling for LLMs before those terms were well defined. The platform was one of the first to allow LLMs to do things like search the web, call APIs or interact with databases, enabling developers to build AI apps even in the early days. Since then they've added observability evals and monitoring through their closed source Lang Smith platform. Now, TechCrunch noted just how many unicorns are emerging from the ecosystem of AI startups. According to them, there were 36 AI unicorns minted in the first half of this year. And Y Combinator has said that they believe that 300 unicorns were created across the entire SaaS boom, a figure that looks like it will be exceeded by AI by the end of the year. One more investment to profile today, although of a very different sort. Meta is making a multibillion dollar bet on AI wearables. Zuckshop has bought a minority stake in Ray ban maker SLR Luxottica. Sources say they purchased around 3% of the company for 3.5 billion. Those sources also say that Meta is considering adding to that investment, bringing their stake to 5% over time. Now the investment certainly cements the idea that smart glasses are a key Meta device play for the AI era They've been offering the Meta Ray Bans for almost four years and recently launched a pair of Oakley branded smart glasses, which is another make from the same parent company. Essler. Luxottica is the largest eyewear company in the world, so the partnership gives Meta access to industry leading manufacturing and distribution networks. Now, leaning this hard into AI devices is a reversal from how Meta addressed the smartphone era. Chamath Palihapitiya, who served as a senior executive back then, has said he pushed the company to develop their own phone in 2007 following the release of the iPhone. That project was never completed, and Facebook was forced to develop unrival hardware for the entire tech cycle. Zuckerberg later said that this was one of his biggest regrets, meaning that his company didn't get to shape the way mobile platforms developed. He is clearly not looking to make the same mistake twice. With Meta entering the AI era with the most established platform launched well ahead of its relevance. Then again, the tech is quickly catching up to make Smart Glasses a functional and ubiquitous AI platform. The International Data Corporation expects sales of Smart Glasses to grow by 47% each year through 2029. So clearly with this deal, Meta is looking to lock in their early dominance of the category. Now, I'm not sure that I think that this will be Meta's only play in the AI hardware space, but it's certainly an unexpected beachhead that has already paid some amount of dividends for them. With that, though, we will wrap the headlines. Next up the Main Episode this episode is brought to you by Blitzy. If you're a technology leader, here's something that probably sounds familiar. Your organization's competitive edge is buried in legacy code that desperately needs modernization. But the resources required feel out of reach. That was the case for a global investment analysis firm. They needed to migrate 70,000 lines of complex Matlab financial algorithms to Python, algorithms that drive investment decisions for trillions in assets. Their estimate? Months of high cost, specialized engineering work. Instead, they partnered with Blitzi. Blitzi's autonomous AI preserved mathematical precision and generated over 80% of the new code base, completing the migration with just five days of engineering time. They cut the timeline by 95% and saved 880 engineering hours. If your organization is facing similar modernization challenges, visit blitzi.com to schedule a consultation and discover how AI powered development can transform your technical capabilities. Today's episode is brought to you by Plum. You put in the hours testing the prompts, refining JSON, and wrangling nodes on the canvas. Now it's time to get paid for it. Plum is the only platform designed for technical creators who want to productize their AI workflows. With Plum, you can build, share, and monetize your flows without giving away your prompts or configuration. When you're ready to make improvements, you can push updates to your subscribers with a single click. Launch your first paid workflow@useplum.com, that's Plum with a B, and start scaling your impact. Today's episode is brought to you by Vanta In Today's Business landscape. Businesses can't just claim security, they have to prove it. Achieving compliance with a framework like SoC2, ISO 27001, HIPAA, GDPR and more is how businesses can demonstrate strong security practices. The problem is that navigating security and compliance is time consuming and complicated. It can take months of work and use up valuable time and resources. Vanta makes it easy and faster by automating compliance across 35 plus frameworks. It gets you audit ready in weeks instead of months and saves you up to 85% of associated costs. In fact, a recent IDC white paper found that Vanta customers achieve $535,000 per year in benefits and the platform pays for itself in just three months. The proof is in the numbers. More than 10,000 global companies trust Vanta for a limited time. Listeners get $1,000 off@vanta.com NLW that's V A N T A.com NLW for $1,000 off we'll go back to the AI Daily Brief. Today we are talking about one of potentially the next big things to hit AI, which is of course GPT5. There is a growing sense that this model is right around the corner and that it may be very significant. So today we're going to talk about everything that we know about it, how it might impact the AI space and yes, a little tea leave speculation around when we might actually get the thing. Now GPT5 has been coming soon for almost a year at this point, but the model that finally gets released will be very different from where it started. In the middle of last year rumors started circulating about the next big models from OpenAI, codenamed Orion. Rumors in the fall were that the model was going to come as early as December, but it wasn't too long after that in around November that we started hearing about industry wide issues around training, with a building narrative that pre training had hit a wall. What we ultimately got in the place of a new flagship model in the series of GPT3, 3.5 and 4 was instead OpenAI's first reasoning model, first O1 and then O3. Now, as we've previously discussed, the launch of reasoning models, it turns out, were a huge inflection point in the adoption of AI. A whole array of new use cases came online, enterprise adoption started going up significantly and the new era of agentix became a viable possibility rather than just something for the future. Now, in and around the reasoning models we did get a GPT4.5. It clearly wasn't a big enough leap to warrant the title, and while the model was a niche hit among people who wanted a better LLM for writing it failed to capture much attention or usage in the broader public. In fact, the model is actually being sunset in the API next week. SWIX from Latent Space had this interesting observation saying I think the quote unquote failure of GPT 4.5 relative to 0.3&04 mini is actually fantastic validation of the bet on the reasoning paradigm. 10 to 100x larger model consistently lost out to smaller reasoning model for an important open domain task and still hold aside what's great about the reasoning models. The point is that there hasn't been a new OpenAI flagship in the series of GPTs for about a year now, and it's been over two years since they felt confident enough to herald a new generation of models with a full numerical upgrade from GPT4. At the same time, the industry has changed quite a bit since the last time we got excited about GPT5 last fall. As we just discussed, the narrative was that scaling hit a wall and people were really wondering if there was any more juice left to squeeze when it came to pre training or if we were just in the era of new methodologies like test time, compute, the reasoning era Agentix, with tool use being the vectors for improvement going forward. At this point now, six to nine months later, there is a pervasive sense of optimism that AI is nowhere near its zenith and that there are still plenty of avenues to pursue in order to improve the technology. Logan Kilpatrick, the head of product at Google AI Studio, recently posted. The next six months of AI are likely to be the most wild we have seen so far. Everything keeps scaling up. More hardware, more model progress, more product knowledge, more AI momentum, more product market fit. Now Logan went to pains to say that this wasn't inspired by any specific thing. He said, it was just inspired by me feeling the progress and then looking out over the next six months and being reminded it's going to continue. So bringing it back to GPT5, what part of that excitement can we attribute to this new model? The main expectation around GPT5 is that it will represent a unification of OpenAI's technology. In a recent podcast, OpenAI's head of developer experience Romaine Hewitt said, we're truly excited not to just make a net new great frontier model. We're also going to unify our two series. The breakthrough of reasoning in the O series and the breakthroughs in multimodality in the GPT series will be unified and that will be GPT5. One of the core promises is that GPT5 will do away with model switching. In a Reddit ama from May, OpenAI VP Jerry Twarak wrote, GPT5 is our next foundation model that is meant to just make everything our models can currently do better and with less model switching. Interestingly, he also referred to their operator agent as a quote product surface, suggesting perhaps that GPT5 will feature tighter integration with agentic tools. Back in February, Sam Altman had said something similar about the model switching idea. In a post on X, he wrote, we want AI to just work for you. We realize how complicated our model and product offerings have gotten. We hate the model picker as much as you do and want to return to magic unified intelligence. Now this post was a couple weeks before GPT5 and he said we will next ship GPT 4.5, the model we called Orion internally as our last non chain of thought model. After that a top goal for us is to unify O series models and GPT series models by creating systems that can use all of our tools, know when to think for a long time or not, and generally be useful for a wide range of tasks. Now getting a little bit more up to date, developer Buijin Ngoc compiled some other nuggets of information gleaned from recent interviews and leaks. He expects a 256k context window, which would put GPT5 roughly in line with most competitors, but not as large as Google's million token context window. Multimodality likely means we'll see native video, image and audio inputs, and perhaps even outputs. Another current rumor is that OpenAI will adopt the mixture of experts architecture brought to prominence recently by several Chinese labs. This architecture means that only part of the model is engaged at a time, allowing for a higher number of total parameters while keeping inference costs down. There are estimates that inference costs could be 60% lower per token than GPT4. O memory is another factor that could see improvement, which would lead to more performant agent operation as well. Not commented for builders, this likely means you have to rethink prompt design for giant context, expect richer tool calling that mixes text with time based media and budget for lower latency cheaper API calls. Despite a larger model, GPT5 looks less like a parameter bump and more like a systems integration milestone, folding multiple specialized capabilities into one cohesive model. And frankly, even without the rumors, that seems like a reasonable assumption. Since its release, OpenAI has been tacking features on to GPT4, including things like memory and updated image generation. GPT5 represents their first opportunity to bake these features in natively, allowing the model to be trained from the ground up on how to make best use of the tools it has access to. RasRx commented GPT5 might be the first model that feels like true AGI if OpenAI integrates full O4 or 04 Pro reasoning plus agentic tool use within the chain of thought, operator codex and deep research, we're talking about a model that can think, plan, act and adapt like never before. AGI Vibes Incoming Now One of the reasons we think that GPT5 is nearing release is that users are starting to report seeing a ton of AB testing of something new on the platform. Specifically, it appears that OpenAI is testing how reasoning traces are presented. One X user, for example, suggested that a recent interaction they had looked like hybrid reasoning in response to a prompt using the 4o model, which again, is not a natively reasoning model. Chatgpt said, Just a moment, I want to give this one the extra thought it deserves. That certainly would imply that OpenAI has figured out a way to differentiate between the prompts that should use reasoning and those that shouldn't. Another interesting feature on that post is a button labeled Answer Now. That presumably cuts the reasoning short and forces the model to just spit out an answer based on how much thinking it's done at that point. One of the problems with hybrid reasoning is that if the model starts thinking and goes down some rabbit hole, there hasn't historically been a good way to make it stop. So again, while it's just a guess, it seems like Answer now might be a UX change to address that issue. So if the idea is that GPT5 will natively bring together all of these features that have been bolted on, perhaps we also need a slightly different way of thinking about what an LLM actually is. Karina Nguyen, a researcher and product staffer at OpenAI, recently posted. Super Intelligent Operating System Believe it or not, she is not talking about my startup. Instead, cryptically, it sounds like it's sort of what's being described in the rumors of GPT5. Rather than treating it like an individual model, you interact with, GPT5 kind of sounds like an AI operating system, TJ Ridgeway writes. So GPT5 is the AGI framework. This is why they're all pivoting to superintelligence discussion. The one ring to rule them all. In response to the idea that superintelligence won't just be a scaled LLM, he added, yes, I pretty much agree. What I'm saying here is this roadmap they put out is the framework through which AGI will be achieved, not AGI itself. I believe innovations in long term memory are also a crucial aspect that is just now being explored. It is worth noting that back in January Sam Altman did mention quote, we're now confident we know how to build AGI as we have traditionally understood it. Still, a practical question after all of this is whether any of it will be enough to amaze heavy AI users specifically. As much as we're summing up all of these different rumors here, none of them really suggest, at least not yet, a massive change in capability or anything particularly new. Instead, these rumors all focus on bringing everything together and making the user experience more seamless. Chubby, for example, wrote, for hardcore users, GPT5 will be a bit of a disappointment if the rumors are to be believed. Rumor has it that Sam Altman is not particularly impressed with the performance and improvements. Compared to older models such as GPT4 0 and 03, GPT5 is more of an iterative improvement that certainly shows significant leaps in benchmarks. But compared to benchmarks such as reasoning at the end of last year 01 or deep research at the beginning of this year, GPT5 is more of the same, just slightly better. In this respect, GPT5 is probably not the qualitative leap that hardcore users had hoped for. And yet, Chubby wrote, maybe the hardcore users are not the point they continue. For the vast majority, however, GPT5 will be a quantum leap. When I talk to friends, I almost always hear the same thing. ChatGPT is great. They say it will help them get more out of their university studies, answer all their questions, and even provide excellent advice on medical matters. When asked which model they would use for this, the answer is always the same. GPT4O. Of course, they either don't know anything about O3 or due to the complicated nomenclature, they consider O3 to be the inferior model because it's older, because 3 becomes before 4. Some people think that simply. However, GPT5 will be an all in one model. Depending on the request, the appropriate amount of inference will be applied and reasoning will be carried out. This means that all those who previously used only GPT4O will suddenly receive much better answers with GPT5 than before because they did not use or were not aware of the full strength and range of ChatGPT's model capacity. Now this I think, is a super important point. The vast majority of ChatGPT users, even at this stage, are not subscribers. They haven't used Deep research and they don't understand what a reasoning model is, or at least how it's different. In other words, removing ChatGPT's model selector isn't just a minor UX improvement. Instead, it fundamentally will broaden the average user's experience by making all of the myriad features available to them without them having to know about what those features actually do. Think back to the Deep SEQ moment at the beginning of the year. The viral breakthrough was not that there was better reasoning. The full version of 01 had already been out for 2 months and was clearly the better model. The innovation was simply putting reasoning right in front of users who had never experienced it before. With the full reasoning traces on display, the average comment about deepseek wasn't just that it was powerful, it's that it was really cute or cool because it talked to itself before responding. OpenAI then has the opportunity to give the average inexperienced user that type of moment of delight with GPT5. Even if every single feature already exists and the performance bump is minor, improved accessibility is a huge deal for the average GPT5 user. And yet I do think that unfortunately for OpenAI, that still might not be enough. Or at least not for long. It's very clear that in the wake of Mark Zuckerberg's poaching spree, the stakes for OpenAI are building. Zuck's superintelligence team is now largely in place, and then it's worth then trying to speculate on what their big play is. In the lead up to the release of Llama 4 during the spring, Zuckerberg set off on a media tour, discussing his AI plans at length. There were really two big themes. The first was making an automated advertising platform, and the second was introducing AI friends to meta social media platforms. Hold aside what you think of those ideas, it is pretty undeniable that they're both fairly modest ambitions. In other words, neither is something you would necessarily want to spend hundreds of millions of dollars in payroll to achieve. And so whether Zuckerberg is aiming for a straight shot super intelligence play or iterative model releases that are actually keeping up with and pushing the state of the art, he clearly has something much larger than just better advertising automation in mind. One observation that many have made is that Zuckerberg has put together a team of experts with a wide range of skills. He poached the reasoning team lead from OpenAI, a multimodal expert from Google, an Edge model developer from Apple, and the list goes on. In other words, a full Stack team capable of recreating everyone OpenAI or anyone else has on offer from scratch. Point being, whatever they're building, it's not AI friends. Altman says he's not concerned. In an interview at the Sun Valley Conference, he was asked how he's feeling about the talent war and responded, fine. Good. We have obviously an incredibly talented team and I think they really love what they're doing. Obviously some people will go to different places. There's a lot of excitement, I guess you could say in the industry. But no, I think we feel fine. At the same time, it is pretty clearly undeniable that whether Sam is a part of this or not, OpenAI leadership in general is starting to feel the heat. We've seen changes in compensation packages, memos that suggest the feeling was of having their house broken into. And so for this reason, I think that the GPT5 moment, whether OpenAI wants it to be or not, is going to be seen as hugely significant and reflective of the state of play when it comes to the broader industry. Back in April, even before this aggressive talent war, Altman tweeted, change of plans. We are going to release O3 and 04 Mini after all, probably in a couple of weeks and then do GPT5 in a few months. There are a bunch of reasons for this, but the most exciting one is that we are going to be able to make GPT5 much better than we originally thought. We also found it harder than we thought it was going to be to smoothly integrate everything, and we want to make sure we have enough capacity to support what we expect to be unprecedented demand. So just how soon is this thing coming? Ultimately, that among all of this is the biggest rumor. We have gotten a bunch of hints recently from OpenAI insiders that something big is coming in the next week or two. But I also think that OpenAI knows the stakes of GPT5. I don't think they're feeling so much pressure that they're going to release something that is anything less than extremely impressive. Still, maybe we have an exciting Midsummer tree coming up. Certainly the chorus of rumors is getting louder and as they get more credible, I will be sure to let you know it here. For now, though, that is going to do it for today's AI Daily Brief. Until next time, peace.
