Transcript
A (0:00)
Google just ran a three layer strategy move in one day at their Cloud Next conference they announced new TPU chips. Chrome is turning into an AI coworker and a multi billion dollar compute deal with Miriam Ratty, the former co founder of OpenAI, her company thinking Machine Labs. I think this is one of the clearest signals from me so far that Google is structurally perhaps ahead of OpenAI and Amazon in the AI stack. Before that though, I want to talk about the fact that OpenAI is teaming up with Infosys to get ChatGPT into 60 plus countries of enterprise deals. Bloomberg reports an unauthorized group breached Anthropic's new cyber tool Mythos. And there is a new research lab called Neocognition that has landed $40 million to build agents that actually specialize like humans. So we're going to get into all of that on the show today. Our first story is that 10x science, this is a Stanford spin out of a Nobel laureate Carolyn Bertozzi's lab. They just closed a $4.8 million seed which was led by Initialized Capital. What they're doing that's so fascinating to me is that basically there's this problem where models like DeepMind's protein predictor, they're spitting out thousands of drug candidates. So there's, there's just thousands and thousands of these drug candidates. And there's a huge bottleneck in pharma where it's not just about, you know, getting all of these candidates, but it's actually triaging them. It's actually figuring out like which of all the candidates is worth pursuing to try to make medicine or therapeutics. And basically the standard triage tool is just mass spectrometry. So it's very slow, it's very hard to interpret. It's, you know, domain experts only. 10x science is basically just building a SaaS layer on top of that. And they have deterministic chemistry plus AI agents to try and make the analysis traceable and explainable, which basically matters because regulators don't accept a black box answer on what the molecule does, right? So just because an AI model's like, oh, oh my gosh, we discovered this super cool thing like this isn't going to fly. Regulators don't want that. There's no, we can't have the black box on the molecules. So we have to be able to fully understand it, fully test it. And knowing which of these, which of these chemicals, which of these drugs to test is a big problem. So everyone in the AI biotech conversation is basically talking about the generative side. And almost nobody is building the kind of picks and shovels layer that's underneath of it. And so this is why I think 10x science is an interesting company. The next thing I want to talk about is Neocognition. So this is an AI research lab that came out of stealth with about $40 million in their seed round. Yu sue, an Ohio State professor who runs an AI agent lab there, is the founder. And the round was led by Cambium Capital and Walden Catalyst Ventures. Intel CEO Lit Bootang and also Databricks co founder Ayaan Stoka both wrote angel checks into this. So I think that's really strong signal right when you, when you have the CEO of Intel and when you have kind of these high profile co founders of something like databricks, this is phenomenal. But essentially the thesis on this company that I think is important is that the current AI agents succeed maybe 50% of the time because they're basically unreliable generalists. And so what they're arguing is that humans aren't great at doing tasks just because we know everything. We're great because we specialize fast when we're dropped into a new domain. So Neocognition is trying to build agents that self specialize the same way. Instead of kind of the current model where you handle, you know, craft a custom agent for every vertical. I've built enough custom agent workflows to know that kind of this per vertical approach doesn't really scale. You run out of engineers before you run out of use cases. So if Neocognition can actually ship an agent that learns the rules of a new environment and if it's doing this on its own, which I think is definitely going to be the key, I think that is a massive win. Right now they have only 15 people on their team. It's mostly PhDs. Definitely an early company, but I think this is one that's worth watching. Okay, so we have some bad news for Anthropic today. Uh, there is a report from Bloomberg yesterday that an unauthorized group got access to Mythos, which is Anthropic's new exclusive enterprise cybersecurity AI tool that's going to take over the world. Right. They gave it to a handful of enterprises because it's so dangerous and good at getting, you know, finding security vulnerabilities. Um, I think this is probably not good news because they, you know, they, they made a lot of hype about how dangerous it was and they only gave it to special people. And if it really is getting leaked that's not great for them. I think what basically happened, if you, if you read the reporting on Bloomberg, there was a group that operates out of a private discord channel focused on unreleased AI models and they got credentials from a third party contractor working with Anthropic. They then guessed the model's endpoint URL based on the pattern that Anthropic uses for other models. And that's it. They then gave Bloomberg screenshots and a live demo to prove access. Anthropic's response to all of this is that they're investigating and so far they say they found no evidence that it touched Anthropic's own systems. No, just the vendor environment. I think this is interesting because this is kind of where a lot of these security incidents are happening right now. It's not really a model exploit, it's just a contractor credential plus kind of a predictable URL pattern. And my advice that I would give to people is that if you are a company that is running AI tools, make sure that you are auditing who is on your vendor side and who has access to all those tools right now. Kind of the bigger thing is that this is also happening in other places. Anthropic is actively in early talks with Goldman Sachs, JP Morgan, Morgan Stanley about an October ipo. So any sort of bad PR story for Anthropic right now is not going to be great. They definitely want to get on top of this. Next up, I have some OpenAI news. OpenAI and Infosys just announced a deal to push ChatGPT and specifically Codex, their coding tool. It's kind of a competitor to cloud code into Infosys enterprise client base. This is across more than 60 countries. They don't, they didn't really disclose the terms or how much money is changing hands in this. But TechCrunch did report on this earlier today and basically for context on Infosys, they did about $267 million in AI service revenue last quarter, about 55% of their total. And they play in kind of the massive enterprise transformation deals where Microsoft and Accenture often take the system integrator slot. So right now infosys is getting OpenAI models baked into their Topaz AI platform and, and OpenAI gets a channel into Fortune tier accounts that they don't reach directly. I think OpenAI is trying to make a big push for enterprise right now. I mean, obviously this has been a goal for a long time, but this big push is coming right now as it feels like Anthropic is really running away with the market in enterprise. I've worked with a lot of enterprises that bought the ChatGPT Enterprise licenses and they kind of sat on them for like a year because they didn't wire them into anything real. This is basically the gap that I think Infosys is trying to fill. And strategically I think this is kind of mirroring what Microsoft already does by bundling Azure OpenAI with Copilot. Except Infosys plays in a lot of deals where Microsoft doesn't always win. It's not, it's not in the deals, it doesn't have the distribution. So it's kind of a new channel. I would expect to see a deal like this to help push OpenAI's revenue higher. OpenAI obviously wants, you know, enterprise revenue to catch up to Anthropic, which is now over $30 billion in annualized revenue. They need this type of distribution in order to try to catch up to Anthropic. Okay, if you're still paying for chat, GPT, Claude, Gemini and Grok by the way, 11 labs for audio and maybe like separate image generators. I have to tell you about AI Box. This is what I personally have built. It's what I recommend to my friends who keep asking me how to actually use AI without going, you know, without spending a ton of money on subscriptions. Basically a box gives you access to over 80 AI models in one place. All of the top models. So you can pick whatever one is best for the task that you're working on or just ask a question and we'll pick the best model for it. The part that I think is super useful is our automation builder. You could just describe a tool that you want in plain English and it will build a workflow for you. So you don't need to code, you don't need to wrestle with a new platform. You just describe it and it will build it. It's 8.99amonth, very cheap. I think this saves you a ton of money on just all of the different subscription platforms and you get your hands on every single AI model that is worth using. We have a ton of news from Google. Their Google Cloud Next conference kicked off today and Google did not come empty handed. They had three major announcements. New TPUs Chrome is an AI coworker and a thinking machine Labs compute deal and I think taken together basically they're going to make Google the most structurally positioned player in the AI stack right now. I'm going to walk through all of these different deals, but yeah, I've been Pretty impressed by their announcements and of course we could see other players, but I would definitely say never count Google out in this industry. So first the chips deal. Google announced two new TPUs. They have the TPU8T for training and the TPU8I for for inference. The split alone is, I think, what's really interesting to me. Inference is the dominant cost of running AI and production right now and having these kind of dedicated inference silicon matters economically. It makes a big impact to the bottom line. So Google's claims is that it is three times faster at training and 80% better at performance per dollar against the Nvidia alternatives and the ability to scale more than a million TPUs in a single cluster. So they're, they're hitting a bunch of different things here, but Nvidia has traditionally. I don't know this is necessarily even Nvidia's fault, although I'm sure some people will disagree with me here. But Nvidia traditionally has had a problem with the cluster sizing. And I remember when OpenAI's or not OpenAI, when Xai did their first like Giga Terminal, I don't know, whatever, they have a robot name for it. So now I can't remember, but they put together a hundred GP Nvidia, 200,000 Nvidia GPUs and it was like this big deal. They kind of figured this out and other people started building these bigger and bigger clusters. Well, and it was kind of like they had to figure out a lot of the infrastructure and build it and think about it. Well, it looks like Google is building straight for that. And the ability to have more than a million TPUs in a single cluster is insane. And I mean insanely scalable. So I think those are Google's own numbers. We don't have any independent benchmarks yet. Google is also not killing the Nvidia partnership. They're reselling Vera Rubin in Google Cloud later this year. So I think basically Google wants both chips and it wants to let customers choose, which is kind of the opposite of what AWS's all in One trainium. Bet is right now I think Google will actually come out ahead on that. The other thing that I think is interesting that Google announced is that they're turning Chrome into an AI coworker. So they have a new feature which is called Auto Browse. Gemini powers it and it's basically running inside of your Chrome for workspace and it reads context across your open tabs and automates actual workplace tasks so it can enter your CRM data. It can compare vendor quotes, it can summarize candidate portfolios, it can write up competitor research. If I'm being honest, it sounds like an interesting tool. It sounds useful and I've actually tried some similar things. Microsoft had something like this in, in Microsoft Edge a while back where like all of your tabs that are open, it can just get data and read and interact with them, which is super useful. Sometimes though you need things that are not, you know, in your tabs. And I think this is where Claude's cowork is still going to keep crushing because they're an app. You give an access to your desktop, it go gets, it can get files, it can download things on your local device. So I know that like open Google isn't trying to directly compete with that. They're all, they're kind of saying, you know, they're just like this human in the loop and they're not kind of letting the agents execute without your approval. And so anyways, I think like they're kind of pitching it as a, as a plus. In my personal opinion it's behind the ball. It's kind of not as good as something like Claude Cowork, which actually sometimes I don't want to have to get 100 pop ups asking. I just wanted to get the job done, especially when I've seen it do it like three or four times successfully. I don't want to approve it, I want it to automate it. And so maybe they're playing it kind of a different target audience there. But I still would probably use Cloud Cowork over that. You can also something that is very useful though and kind of like Cloud Cowork, you can save frequent workflows as skills and you can fire them with a forward slash. This is rolling out in the US Workspace. Users are getting this first and there's also a commitment that enterprise prompts are not going to be used to train Google's models, which I think is the right call. A lot of enterprises would be very concerned especially if they wanted to put confidential data in there. But again, and not to like, I don't know, say all of the downsides to this again though, you have to go and fire off a skill manually where something like Cloud Cowork you can just set these things up to automatically run. Okay. The next thing that I thought was interesting is that Google Cloud has just signed a multi billion dollar deal with Miriam Moratti's Thinking Machine Labs. This is a single digit billions. It gives Thinking Machine Labs access to Nvidia's GB300 system on Google Cloud training and deployment services for their first product which is called Tinker, a tool for building custom frontier models. Thinking Machines is raising at a $12 billion valuation and they run a heavy reinforcement learning workload which is basically very insanely compute intensive. Mile ott One of the founding researchers is basically talked a ton about how much Google got them up and running really quickly. So I think this is something that's going to be helpful for their company. The thing that I think ties all of these together is that Google is running a three layer strategy. The bottom layer silicone TPU plus these resold Nvidia's and then they don' which chip wins. I mean obviously they kind of want their own to win but they're going to make money either way even if you buy Nvidia from them. The middle layer is being the compute host for Frontier Labs. Anthropic already runs on TPUs and now thinking Machine Labs is going to be running on Google Cloud. The top layer is the agent layer in the browser and the workspace. Google is, you know, the only player that's hitting all three credibly right now. Anthropic is obviously crushing it with end users and you know, kind of in that top layer but they're not making their own chips, right. And so Google has a big advantage when they own the entire stack. Mike Microsoft is really heavy on OpenAI at layer 2 but really weak on Silicone. Amazon is betting the farm on Anthropic. Nvidia is the chip layer and doesn't touch the application layer. Google is basically the only company really doing the full stack. So I think the strongest argument against this is that Gemini still hasn't closed the gap with GPT 5.4 or Claude Opus 4.7 on the benchmarks that consumers actually care about. And so I think, you know, maybe those TPU numbers are Google's own. And you could also say, you know, until like an independent lab publishes real world training runs on the 8T and on the 8I the 3X claim that they've given is marketing. And I mean I don't want to, I don't want to doubt Google too much, but you have to be, you have to be fair. There is a lot of these AI companies and Google's been known to do this in the past to make big claims on marketing and it doesn't always stack up in the real world. But here's why I still think even with all of that, Google is in an incredible position right now. The customer really doesn't care whether Gemini tops the ELO leaderboards, if they can run inference cheaper on Google Stack and if they can serve it to their employees through Chrome and wire it through their workspace data without, you know, having this big huge system integrator structural position. I think matters way more than just pure benchmark lead in the next 12 months and Google is integrated with every tool that people are using right now and so I think Google's going to do quite well. The other thing that I think is interesting is that Google is doing this with a DOJ antitrust kind of suit overhanging them right now. The search remedies case that they've been in for a while now is still live, and turning Chrome into an agent layer that pulls from workspace data across your browser is basically the exact kind of move that makes regulators pay attention. I think you would be wise to watch for some comments coming from the DOJ or the EU in the next few weeks. This is going to draw scrutiny that Microsoft and kind of their OpenAI deal has already been dealing with for a year. What I'm watching personally is the independent TPU8T and 8i benchmarks, whether auto browse graduates from pilots to real rollouts inside of enterprises, and whether more frontier labs migrate to Google Cloud now that thinking machine is in and Anthropic is already there. So I think if even one more lab follows, that is a huge structural shift. All right, that's it for the show today. If you've got something out of this, it would mean the world to me if you could drop a comment on Apple podcast or if you could hit the stars over on Spotify. It's the about tab on Spotify where you can leave a review, leave some stars, drop a comment on Apple. It helps the show a ton, right? I don't ask for it for nothing. It really helps more people find it. It helps in the algorithm. Also, if you want to get access to over 80 models with an automation builder that runs in plain English for 8.99amonth, go check out AI box. AI. There's a link in the description. In any case, I'll catch you guys all in the next episode.
