
Loading summary
Paul Raitzer
I do believe that organizations are going to start to filter their employees and do their annual evaluations or whenever it's happening, they're doing these evaluations. I do believe that AI literacy and competency are going to become a key filter for who stays and who goes. Welcome to the Artificial Intelligence show, the podcast that helps your business grow smarter by making AI approachable and actionable. My name is Paul Raitzer. I'm the founder and CEO of SmartRx and marketing AI institute and I'm your host. Each week I'm joined by my co host and Marketing AI Institute Chief Content Officer Mike Kaput, as we break down all the AI news that matters and give you insights and perspectives that you can use to advance your company and your career. Join us as we accelerate AI literacy for all. Welcome to episode 144 of the Artificial Intelligence Show. I'm your host Paul Raetzer, along with my co host Mike Kaput. We are back in Cleveland after. Well, I had a full week in Vegas, Mike. You were. Got two full days or whatever it was. Yeah, so we were in Vegas for the Google Cloud Next conference last week as well as a couple of other events and programming items. It was a long time to be in Vegas. It was an incredible event. But five nights in Vegas is a lot, man. So yeah, so we're back. It was a crazy week in, in an average week, I think I'll, I've said this before, like I'll Curate maybe like 30 to 40 links that Mike then goes through and like, you know, picks the things for the show. I was well north of 50 last week and, and we, I mean we literally just cut stuff as we were getting ready to come on here. So it, it is a lot and I, I expect this week is going to maybe be on par with last week. So got a ton to cover. We're going to go kind of rapid fire style through almost all of these. There's, there's a couple will linger on a little bit more, but it's not going to be the, the standard like three main topics and then the rest. There was just so much to get through. We're going to try and move through things pretty quickly, but. All right, so before we get started, this episode is brought to us by the Marketing AI Conference or Macon. You can learn more at Macon AI that is M A I C. This is our 6th annual in person conference. It's the flagship event for Marketing AI Institute. It's taking place October 14th to the 16th in Cleveland, Ohio at the Convention center right across from the Rock and Roll hall of Fame and Cleveland Brown Stadium, at least for now until the Browns may be moving a couple years. So we announced the first 19 speakers. There's plenty more announcements coming, but you can check out the agenda. The preliminary. No, we didn't put the full agenda up yet. We still got last year's agenda, I think is the example. Right?
Mike Kaput
Yeah. But we do have, we have some speakers.
Paul Raitzer
Okay. Yeah. So you can go to the speaker page, see the speakers, and we'll be updating the the agenda soon. You can look at the four workshops. So there's optional workshops on October 14th, half day workshops on that day. You can go check those four out again. It is October 14th to the 16th in Cleveland. Prices go up on April 26th. You got a little time to take advantage of early bird pricing and you can go again to Macon AI to learn more about that. I am really looking forward to this year. It's going to be the best year yet. We've got some pretty cool things planned. Even just like last week I was meeting with the team to talk about the, you know, the exhibit hall and some of the things we're planning on doing there took a little inspiration from the Google Cloud Next event. So yeah, check that out and we'd love to see you in Cleveland in October.
Mike Kaput
All right, Paul, so let's dive into it. Our first topic that has happened in the past week is that ChatGPT can now, if you so choose, remember and reference all of your past conversations with it, not just the ones you have explicitly saved, thanks to a new feature release from OpenAI. So what can happen now? Here is that ChatGPT can reference previous chats to deliver more personalized responses, no matter what you happen to be prompting ChatGPT to do. Now this is different from and builds on top of the memory features released last year and goes a bit further. So you know, there used to be and still are those kind of saved memories you can ask ChatGPT to explicitly keep. But now there's also, if you so choose chat history references, which are insights it passively gathers to make future conversations smoother and more relevant. So this is a big step towards OpenAI's long term vision of AI that grows alongside you and one that, as CEO Sam Altman put it, quote, gets to know you over your life. This feature is rolling out now to pro and plus users, except it is not at the moment in the EU and a few other countries due to AI regulations there. And you can still opt out of this entirely or start A temporary chat session if you prefer for your conversation not to be referenced or remembered. So, Paul, this feature is a pretty big deal. I mean, some thoughts I'm wondering if you could kind of respond to here. It seems really, really valuable if you are comfortable with giving chatgpt that kind of access to your past chats and conversations. Not to mention, it seems like if this is something you find tons of value in, you are now probably, I would guess, much less likely to leave ChatGPT if it has all this context and memory about you. But on the other hand, I can also see there are a fair amount of people that appear hesitant here to use this feature.
Paul Raitzer
Yeah, I think that they're both true. Like, it is a very powerful feature. It could be a potentially sticky feature. It's also a feature Google has and is going to build on. And if you think about the data Google has, not only, you know, does it have chat potentially if you use Gemini, but it has emails and search and all these other components that it can know you even more deeply. Google Maps, Waymo, like Start, you know, mixing in all of these other things, your Pixel devices, your Nest cams. So Google, if again you allow it. This is the kind of next frontier these companies are competing for, is to take all of your personal data to create truly personalized experiences for you through your AI, so they can build truly personalized assistance. Apple, if they ever get their selves together, could play in this world too. They're just much more protective of your private information. So it's helpful maybe to highlight a couple of things. There's a great or very helpful memory FAQ that OpenAI also updated related to this. So I would recommend people read this because I think it's very important people understand what their privacy rights are here and what is maybe automatically turned on and what you need to turn off. So from the FAQ page it says ChatGPT can now remember useful details between chats, making its responses more personalized and relevant. As you chat with ChatGPT, whether you're typing, talking, or asking it to generate an image, this is multimodal. It will remember helpful context from previous conversations, like your preferences and interests, and use that to tailor its responses. The more you use ChatGPT, the more useful it becomes. You start to notice improvements over time as it builds a better understanding of what works best for you. You can also teach ChatGPT something new by saying it in a chat. For example, quote, remember that I'm a vegetarian when you recommend a recipe to check what ChatGPT remembers, just ask what do you remember about me? I don't know if you've done that yet, Mike, but I did it. It's kind of fascinating. It says you're in control of what ChatGPT remembers. You can delete individual memories, clear specific or all saved memories, or turn memory off entirely in your settings. If you'd like to have a chat without using or updating memory, use the temporary chat. As you mentioned, Mike, temporary chats won't reference memories and won't create new memories. If memory is on and you want to fully remove something chatgpt remembers, you will need to delete both the saved memories and settings and the chat where you originally shared this information. If this sounds to you like people are just going to not change anything and they're just going to accept whatever OpenAI sets as the standard, you are probably correct. I would imagine that of their what now sounds like 6 or 700 million users, the vast majority are never going to touch these settings and they're just going to not even probably know that this is a thing. So that to me is like memory is just going to be a part of the chat experience moving forward. It says Chat GPT Enterprise workspace owners can turn memory on or off for all users. So if you are a workspace memory or workspace user in enterprise, it doesn't say team. I don't know if team doesn't have that option. You don't have a choice here. It's your admin that has the choice for you. A couple other quick, interesting notes they had in the faq. It says, does memory remember sensitive information? It says memory raises important privacy and safety considerations, especially around what type of information should be remembered and how that information is used. It continued, we're taking steps to reduce bias and have trained ChatGPT not to proactively remember sensitive information like health details, unless you explicitly ask it to. We are continuing to prove how the model handles this type of information. You're in control. You can review and delete safe memories, ask ChatGPT what it remembers about you, and delete specific conversations. This one is interesting to me because I have had a health issue recently that I was definitely talking to Chat GPT about and when I was asking for, you know, what it remembers about me, none of those details are there. I think it's real important for people to know they're there. They're just choosing not to surface them to you. So what they're basically saying is like, we remember everything you ask us to. We just classify stuff and extract it. When you ask us about memories basically so it's how good are they at filtering the sensitive and private information is really the question, not whether or not they know it. And then a couple other quick notes. Do you train your models with memories? Says if if you have the improve the model for everyone setting turned on. So if you're not aware of it, there is an option in your settings, I think it's under privacy where you can turn on or off improve the model for everyone. If it's on, they will use your information to train their models. So it says if you have improved the model for everyone setting turned on, we may use content you've shared with ChatGPT, including past chats, saved memories and memories from those chats to help improve our models. Now a couple quick personal notes here. The first thing I thought about when we first started talking about memory is I have two ChatGPT accounts. I have our Chat GPT team account for the company and I have a personal Chat GBT account. I happen to have a Pro license. Personally, I I'm generally pretty good about having business related conversations within my business account and my personal conversations, my personal account. Except I don't always think to like look and see which account is active. I just go in the app and whatever one happens to be active. Something I think would be correct if you only have a single ChatGPT account and you use it for both personal and business reasons. What I learned is you can go in and ask it to segment those memories. So the way I learned this is I actually gave it the like what do you remember about me? Thing and it sort of like mashed together some stuff. So it gave me, it categorized these as professional interests, business and strategic projects, creative and personal interests, preferences and styles. And so what I found when it first outputted is it was mixing memories of things relevant to me and then things I did for someone else. So if I was like demonstrating it to someone, like how to, you know, create a children's book or something, it thinks I'm writing a children's book. And so all these memories are just like mashed together. And it doesn't know the context. But the part that was really weird is when it did this and it mashed together my own stuff and other people's stuff. I said to it, you're mixing memories of things I've done to help others, like a children's book and my own activities and interests. And it replied, great point and thank you for the clarification. Let me separate those more accurately. And it did. It was weird. Like, it all of a sudden knew everything that wasn't for me and then everything that was for me. So my guess is if you use ChatGPT for personal and business and you say, hey, split my memories into personal memories and things that I, I do for business, my guess is it's going to filter them really well. And then it replied back. It's like, want me to revise what I remember based on your own projects and pursuits versus what you've done for others? I said, yes. And then it said, got it. From now on, I'll make a clear distinction and focus on your own work and interests unless you specify otherwise. If you ever want to highlight work you've supported for others, just let me know. Do you want a fresh recap of your core activities and interests? Now, this is wild. Like, it's there. There's more going on here than just, like, information retrieval. There is, like, an understanding of the context of, like, how are you using these things? And again, said it a million times, this is the dumbest form we're ever going to have. Like, these things are only going to get smarter. And the people that choose to allow them to have access to these memories, again, not just OpenAI, but Google and others, they are going to have very different, more powerful experiences with these chatbots. But it's a very slippery slope also. So, yeah, that's just some context. Check your settings, though. It's a good reminder to check your settings and see what you're allowing to be remembered.
Mike Kaput
Yeah, absolutely. Yeah. It's already opened up some really intriguing possibilities, but like you said, it's how comfortable are you going to be taking advantage of this?
Paul Raitzer
Yep.
Mike Kaput
Next up at Shopify, the CEO is sending a very clear message to staff, which boils down to basically, don't ask for more headcount until you've proven AI can't do the job. So in a memo to staff, CEO Toby Lutke laid out a new standard. Before requesting more people or resources, teams at Shopify must show that AI isn't a viable option. He called AI usage a baseline expectation and said that employees are now being evaluated on how well they integrate it into their daily work. This memo was originally internal, but Lutke published it after he heard it was being leaked by the press. So, like many tech companies out there, Shopify has been trimming its costs and trying to increase its efficiencies and investing heavily in AI. They have AI tools that they sell to customers, like their Sidekick chatbot. And now they're trying to turn that same logic inward to the company's own operations. Now. Paul, some quotes that stood out to me from this memo. He said, quote, what we have learned so far is that using AI well is a skill that needs to be carefully learned. By using it a lot, it's just too unlike everything else. He also said, I've seen many of these people approach implausible tasks, ones we wouldn't even have chosen to tackle before with reflex and brilliant usage of AI to get 100x the work done. And he also mentioned that using AI effectively is now a fundamental expectation of everyone at Shopify, and everyone means everyone. This applies to all of us, including me and the executive team. And Paul, I noticed that you also had a post about this which seems pretty interesting to talk about, saying this will be universal across industries by the end of 2025. Can you maybe talk to us about that a little bit?
Paul Raitzer
There was definitely a lot of noteworthy items from this post and it's not long. I mean, it's probably like a thousand words or something. It wasn't a crazy long memo, but yeah. So the one in particular that jumped out to me is this. Before asking for more headcount resources, that teams must demonstrate why they can't get it done with AI. That is the right approach. Like, so anyone who listen to this podcast rightly knows how pro human I am in all of this. That we have to reskill and upskill is a top priority. We have to try and create opportunities for people as jobs start to become impacted by AI. But as a business leader, that is fundamentally absolutely what you need to be asking of your team. If you want three more customer service reps, first show me why we can't do what they do with AI. If you want two more BDRs, first show me why we can't do what they do with AI. So before you start adding staff, because it's, it's the only responsible thing to do. Because what happens is if, if we only look at today and we say, okay, let's add those customer, you know, service managers or whatever the role is, and then six months from now we realize, oh wait, that's only like a half fte now we don't need those people anymore, then you're in a tough spot. So I think that organizations have a responsibility to maintain as many workers as possible and to rescale and upskill them. But you also have a responsibility to be looking out, you know, 6, 12, 18 months from now and saying, do we really need to make this higher because it's way better to not make the hire than to be in a position to cut that role, you know, in six to 12 months. So I do think that AI forward companies, the ones that have leaders who understand what these models are capable of today and what they're going to be capable of in the next six to 12 months, that is, that is absolutely what they should be doing. And I think we're going to start to see more leaders take a very direct approach to this and be more specific about we are going to require AI usage. I think that, and I'm saying, I think this, I can also verify, I've had these conversations with leaders in the last two weeks that are doing this exact thing, which is AI literacy and competency are going to become a filter for your employment, meaning you're not going to be there if you don't figure this stuff out. So if you look across organizations and I don't care if it's marketing, sales, HR, finance, legal, whatever it is, in the next 12 to 18 months, I do believe that organizations are going to start to filter their employees and do their annual evaluations or whenever it's happening, they're doing these evaluations. I do believe that AI literacy and competency are going to become a key filter for who stays and who goes. And I think the employees that move forward and prove their ability to drive productivity, efficiency, creativity and impact on revenue and growth, they're, they're going to be in the best position to keep their jobs and thrive because revenue per employee numbers are going to go up, profits in theory go up and those people stand to, to, to benefit greatly. I've said this a couple times but like there's no better time in history to be building a company from scratch because all these KPIs you would look at like a revenue per employee number. So depending on your industry, that number, you know, maybe it's 300,000, 400,000 per employee you target in a software company it might be 6, 700, 800,000. If you're Nvidia, it's like 1 million, 1.2 million per employee. I think those numbers are going to get completely reset. And I can, I can say this from personal experience of building our company in a more efficient way and how I look to our future, I don't, I don't think that it's, it's unrealistic for service companies, knowledge based businesses to be doing closer to the Nvidia numbers than to the standard numbers and, but it takes reimagining what those companies look like. And the way you do that is by building an AI literate, AI competent like AI forward workforce. And if everyone on the team is moving in that same direction and constantly saying is there a smarter way to do what we're doing, processes, workflows, campaigns, tasks, the compound effect of that is going to be insane for organizations that get it. So yeah, I feel like there wasn't much in that memo I would disagree with. Honestly. Like I think he he's saying the stuff that I've been hearing that most executives have been unwilling to say publicly.
Mike Kaput
Somewhat related to this is our next topic about the analytics software company Databox. So recently Databox CEO Pete Caputa recently posted on LinkedIn that the company replaced 80% of its customer support and sales development staff with an AI chatbot and actually improved results by 40%. So it's important to note, as Caputo does multiple times in the post and in the comments, that the company reduced their headcount. Well before adopting AI for this use case, he said he had 40 people on the sales, development and customer support teams, but had to reduce headcount to eight because the anticipated demand they had hired for didn't materialize. Now, well after that, the company then deployed Intercom's finbot, which now resolves about half of all their customer chats instantly. So that freed up human reps to focus on personalized outreach to high potential leads, helping drive more revenue. Now Caputa said that Finbot's customer satisfaction scores can still lag behind humans, but its speed and consistency make up for it. And Databox actually improved its results by feeding it better content like integration specific help docs, and use case forums to handle the long tail of customer questions. Now he also admits the kind of next evolution of this is a bit harder. Giving bots the ability to log into accounts and then troubleshoot things like a human would could present some roadblocks. But he also says 18 months ago he wouldn't have believed how far they could have gotten already with AI. Now Paul, you know Pete, and we're all familiar with Databox having used this software. I'm curious about your thoughts here. Pete himself said he hesitated at times to share this story. There were definitely some negative comments about his decision on this post as well. But also, this is a pretty impressive case study of what's possible.
Paul Raitzer
Yeah, so Pete and I go back a long time. So Pete was the architect of the HubSpot Partner Program. And again, any longtime listeners that would know my agency, my former agency that I sold in 2021 was HubSpot's first partner back in 2007. So Pete and I go back all the way to the origins of the partner program at HubSpot. And then he moved over to Databox. I don't know how long he's been CEO there. It's been a while. I want to say it's like six or seven years maybe. So, yeah, we've, we stayed connected. Pete's a good friend and he was not, I would say, and probably admittedly himself, he was not an early adopter of AI. Like, I remember pushing Pete back in like 2017, 18, 19. I was like, dude, you should be building an AI into the business intelligence platform and here's all the opportunities. And to his credit, like, he came around and now he's like all in on infusing it into their product and obviously now into their business. And, you know, I think he was pretty clear that as you highlighted, Mike, these layoffs weren't because of AI. He didn't reduce the staff, but he, as I was just saying in the previous note, he is reducing the number of new hires by using AI. We just won't need as many people going forward. And so that's the right approach. But I, I do think that we're going to hear a lot more stories like this where the layoffs will have been because of AI. So there's going to be a lot of instances. Now, again, I know these things are happening. I hear firsthand these things are happening. But people aren't saying publicly yet this is why it's happening. But they will. So you will have layoffs because of AI, because you see leaders who look at teams and say, we don't need five people doing that anymore. You will figure it out with two of you because they know that AI is now capable of assisting these different roles. And again, it's across departments, but it's, it's starting in like, marketing is a, is a big one right now. Sales is a big one. And so what's going to happen is you're going to have leaders put constraints on teams and challenge them to achieve new levels of, new levels of efficiency and productivity with AI. And again, I'm not saying this is the right approach. I'm just telling you it's happening now and it's going to happen at a much bigger scale as the year goes on. And as a business leader, there's no greater way to drive innovation than to create deadlines and restraints. So if, if everything is great and like, there's no real limits on budgets and we have as many people as you could possibly need to do whatever you want. People get lazy. And so this is what, and this is like the tech culture. This is driven largely in tech. It is constrained resources and then help people realize what they're actually capable of doing under constrained resources. And so I think given the economy, given a number of other advancements in AI models, I think you're going to see leaders who put constraint resources on their teams and say you can, you can produce greater with fewer people. We, we believe you can now go do it and they're going to challenge them to do it. It's going to be uncomfortable. It, it's not going to be maybe fun to be like in an AI emergent company that has maybe hundreds of people in your marketing department and you don't think you need as many. You have to make some difficult decisions. But it's going to happen. And it's again, it's happening. It's just not being publicly talked about yet. They're what I call quiet AI layoffs that there are, there are layoffs happening that are not being put underneath that headline, but they will be.
Mike Kaput
You know, I think it's really just worth reiterating that this is such a useful piece of AI related career advice too. I would say just as someone like myself who's trying to navigate this is really, if you can take a step back and get some perspective and put pretend you are the CEO of a company. Think about like Pete. Go to Pete's post to start because he goes into the hard parts of his decisions that he had to make. He's like, well, you know, here's what I was thinking. Here's how I thought about it. Not everyone understands the decisions I have to make. Like, that's a good barometer for how you want to be thinking about AI in your own career and function, I think.
Paul Raitzer
Yeah. And the other one, Mike, that you know, came up in some conversations last week is take your top players, take, take the people on your team who are figuring this out or listen to podcasts, taking online course, reading books like they're doing everything to figure it out. And they're pushing like the limits of ChatGPT and Google Gemini and the prompting and they're testing out all these new tools. Those people are going to 10x the value they create on a team. How in the world can you talk to that person in like an annual review when they're looking around, seeing the other people on their team who aren't doing anything. They're not, they don't know how to prompt anything. They're not using ChatGPT. They haven't taken any initiative to learn AI. And now you have someone who's like creating 2, 3, 10x the value and all of their prompts are the ones that everybody else is using and they're the ones that are like educating the rest of the team. How do you talk to that person with a straight face and say, yeah, we're going to keep everybody else around, like we'll eventually they'll figure it out. It's like, no, because now as, as the person on that team who's doing all this work, right? I'm like looking around saying, where can I go where I'm going to be really valued and I'm not going to be pulled down by everybody who's not figuring this out and refuses to like use AI. So I think that that's where you're going to have these AI forward practitioners who want to be around other AI forward practitioners and they don't want to be kept down compensation wise, career trajectory wise. When they're the one who's doing everything that's being asked of them, or maybe not even being asked yet, they're just like the innovator on the team and they start to feel like they're being constrained by their leadership who maybe doesn't understand AI or by their peers who refuse to use it. So again, there's, there's just basic business fundamentals that tell us things are going to change like they have to, or the best talent is going to leave and go somewhere where that talent's appreciated.
Mike Kaput
Our next topic this week is one you've already alluded to, Paul, which was Google Cloud Next 25 just wrapped up this past week in Las Vegas and some of our team members were there for some, or in your case, all of the event. So we wanted to briefly cover some of the top announcements to come out of the event. There were a ton of them. Google literally published a list of 229 announcements in total. So obviously we're not going to go through all those, but you can find the link in the show notes. But just some interesting highlights here. Kind of what's top of mind at Google Cloud is a bunch of announcements centered around Gemini 2.5 Pro, which is now available in public preview. It's Google's most powerful model to date. It tops the Chatbot arena rankings and it's designed for advanced reasoning and coding, among other things. Alongside this were also announcements About a leaner, faster Gemini 2.5 flash, as well as major upgrades to image, audio, video, and even music generation across some of Google's different models, including imagine3, chirp3, vo2 and lyria. Google also announced a handful of important updates to Agent Space, which is its platform that connects your work apps to Google's AI models and agents. So you can use these AI models and agents with all your information and data. And there were a ton of updates about AI infrastructure. Google debuted new GPUs, TPUs, high speed networking and storage optimized for training and inference at scale. So, Paul, this was a really cool event that you and I and some of our team got to experience. I want to kind of get your view on. Were there any big takeaways you had from the event first? And then also I have to have you share your experience on night one, before I got there, when Google used AI to recreate the wizard of Oz in the Las Vegas sphere.
Paul Raitzer
Yeah, the wizard of Oz experience was crazy. So I'd never been to the Sphere. If you aren't familiar with it, look it up. Like, you can watch some YouTube videos of it. It's wild. So I know it's a concert hall, but they also are creating these other experiences. And so Sarah Kennedy, who is a friend of mine and who we collaborate with, who's the VP of global demand and growth marketing at Google Cloud, she sort of spearheaded this event and experience, and it was just remarkable. So what they're doing is James Dolan, who's the CEO of the Sphere, had sort of approached Google about creating this, about re, you know, bringing this 1939 film to life in this amazing arena. And so they've been working with the Google team for, I think, like two years now to do this. And what they did is they didn't show us the full film. It's not ready yet. It comes out August 28th this year. It debuts at the Sphere. But they showed us almost like a documentary of this, this building, of this experience, of taking this film that's basically in a rectangle from 1939 in low resolution and expanding it to fit this massive screen and to create this, you know, multimedia experience with, like, wind blowing up from the floors and the sheet, the seats shaking when, like, thunder hits. It is so crazy. So there was all these innovations and they talked about, like, they interviewed this one guy from Google DeepMind, and they said, like, hey, when this project started, like, what did you think was impossible? And he's like, everything. Like, there was nothing. We were Doing that, the models at that moment could actually achieve.
Mike Kaput
Wow.
Paul Raitzer
And so they had to create all these breakthrough innovations specifically with Gemini vo, which is their video gen model, and Imagine, which is the image generation. And the three components they focused on is one called super resolution, and then in painting and out painting. So what happens? This is from a Google blog post that we'll link to. It says, using versions of veo, Imagine and Gemini specifically tuned for the task, the Google teams and their partners developed an AI based super resolution tool to turn those tiny celluloid. Celluloid, yeah, frames from 1939 into ultra, ultra high definition imagery that will pop inside the sphere. And it does. Having seen it. Then the teams perform AI outpainting to expand the scope of scenes to both fill the space and fill the gaps created by camera cuts and framing limitations. Finally, through performance generation, they're incorporating composites of those frame performances into the expanded environments. Together, these techniques help achieve the natural gestures, staging and fine details that conventional CGI struggles to match. Like there was an example where they showed Dorothy in a scene where she talks to the uncle initially when she comes in the door and then she like goes in to the aunt or whatever. Well, in the tradition in the original film, the uncle's off screen has nothing to do with it. Well, in this expanded version, he's there like they're, they're recreating characters that would have been in a wider shot. They're actually like the AI is creating these characters with natural moves. It was so wild to see. So yeah, there's no documentary about this. Like the Wall Street Journal had an article about it. You can't really go watch video of it. It was a private event, but just remarkable. And it does show. Like the thing that I took away with it was the, like the Human Machine collaboration. Like this. This wasn't like you just gave it to Gemini and Gemini figured all this stuff out. There was dozens of the top Minds within Google, DeepMind and Google Cloud within Google working on this, envisioning this, and then like pushing the limits of the models in many cases creating entirely new techniques to make this possible. And it does. It's one of those moments where like when I first put on like a Vision Pro and you're like, oh, wow, like this product might not take off, but this is a whole new experience. That was what I felt when I was at the sphere. It's like, this is totally. This opens up all kinds of incredible possibilities for things that could be done in that kind of environment with AI working with the human. So yeah, wizard of O thing was nuts again. Go check that out. And then the, the big thing that stood out to me and I, I sat through a lot of sessions and content Agent Space. Like we talked about agent space in December 2024 on episode 127 when they first announced it, I think we just kind of mentioned it because there wasn't much information about it. It wasn't actually available. It was just sort of like a preview thing. It was the most impressive thing maybe outside of the sphere that I saw last week. So basic premise is like, it's a single space that has your prompt galleries, your agent galleries enables you to, in a NOD code environment to build agents to do whatever you want to do. Connects to third party software and data. So it basically becomes this like platform where you live and do everything you need to do. It has deep research built in, it has audio overviews built in. You can turn anything into a presentation on the fly. Create a deep research project. Okay, turn this into 10 PowerPoint slides or Google Sheets or Google Slides that I want to turn, you know, use for my presentation. It's got Notebook LM baked in. So it's like it. The vision for it is powerful and you could see how it becomes like a control panel basically for a knowledge worker to just like all the tools they need are just living right there. It's not available yet though. Like you have to go request access. So that was the only frustrating part is like I don't even know like when it's going to be available, how do we get it? Is it? I have no idea. And I was there and I still don't know. So that's the only downside I guess is it doesn't really exist, that I can tell. But I do know people in Google are using it. So it is a thing. It's just not a publicly available thing yet for most of us. And then the last thing I'll note is I was at a part of a leader circle event and had the privilege of sitting there and listening to Sundar Pichai be interviewed. And he did say that he expects the pace of the model advancements to continue for at least 12 to 18 months. And he said specifically new major models every three to four months. So when we say these models are going to keep getting smarter and the velocity is there, this is the CEO of Alphabet and Google saying like, yes, for at least the next year to year and a half, every three to four months, we are going to see massive model improvements, which is just crazy. To think about.
Mike Kaput
Wow. So our next piece of news here, actually, I believe, dropped while we were recording our last podcast, so we didn't get to cover it then, but has some relevance probably for the next coming week or two. So, on April 4, according to a post on X from OpenAI CEO Sam Altman, OpenAI will in fact be releasing its O3 and O4 mini models after all. So Altman said, quote, change of plans. We are going to release O3 and O4 mini after all, probably in a couple of weeks, and then do GPT5 in a few months. There are a bunch of reasons for this, but the most exciting one is that we're going to be able to make GPT5 much better than we originally thought. We also found it harder than we thought it was going to be to smoothly integrate everything, and we want to make sure we have enough capacity to support what we expect to be unprecedented demand. Now, it Sounds like originally OpenAI was going to shelve a separate launch for these models and instead kind of bake them into GPT5. Kind of the goal here is they want kind of a unified system that has voice, canvas, search, all in a single system that can intelligently decide when to think deeply or give you faster answers. But until that happens, it sounds like we're getting some new models to possibly experiment with very soon. So Paul, what does this kind of pivot mean? Like, is GPT5 behind schedule, not performing as planned, or just changing the approach here?
Paul Raitzer
It might be, who knows, might be a mix of all the above. My guess is again that the reasoning models opened up a whole new scaling law and they've been trying to figure out since the fall when they released the first reasoning model, you know, where, where this goes and when they're going to kind of bring it all together. So I think it's just, you know, things are fluid internally and they're GPU constrained. So they, that might be the biggest thing, is they're limited by their ability to do inference with their chips. And I think that the demand for image generation since it came out a few weeks ago maybe made that even more challenging, where Sam's basically on Twitter begging for chips from somebody, anybody. So that could, that could be. It is, it just might be they're constrained by compute power to do these things. Now, Sam did Tweet on Sunday, April 13, We've got a lot of good stuff for you this coming week, kicking it off tomorrow. And then OpenAI just tweeted about an hour ago this where it's Monday, April 14th right now, Mike and I recording this developers with a handshake emoji super massive black hole live stream 10am Pacific Time so this is one 1pm Eastern Time. So I asked Grok. So if you haven't done this yet, if you have in in Twitter X there's a little Grok symbol which ironically is a black hole. And if you have a tweet like a lot of these times, like the AI people like to vague tweet stuff and you're like what the hell does that mean? Like only you know, a small group people have any clue what they're talking about. So you can just click the little grox symbol and it'll actually like give you some, some hints at maybe what it means. So I asked Grok because I have no idea what the supermassive black hole thing is. And it said OpenAI's cryptic post hints at a major developer focused announcement tying developers with supermassive black holes. I could figure that part out. Possibly symbolizing a groundbreaking AI tool with immense potential. Set for live stream at 1pm which we already talked about. The supermassive black hole reference aligns with recent astronomical discoveries like the awakening black hole. Yeah, that's nothing to do with this. Okay. Community speculation centers on Quasar Alpha, a rumor open AI model with a 1 million token context window, potentially linked due to its name, name's cosmic connection and metadata similarities with OpenAI's API suggesting a stealth release for developers. I'd never heard of Quasar Alpha. I don't know if that's Gro's making that up or if that's actually a thing. Oh wait, here, I'll hit explain Quasar Alpha in in Grok and let's see what happens. Is an intriguing AI model that emerged on the scene in early April 2025 generates significant buzz within the AI and developer communities and then it goes into a whole bunch of details. So yeah, I don't know, maybe Quasar Alpha is coming today, whatever that is.
Mike Kaput
So Sam Altman also took the stage at TED this past week and talked about some interesting points about the present and future of OpenAI. So he actually revealed that ChatGPT's user base has doubled in recent weeks and now it's used heavily by about 10% of the global population. That growth includes 500 million weekly active users. He confirmed OpenAI is working on a powerful open source model quote near the frontier, responding to pressure from challengers like Deep Seat. And he even said, hey, we were late to act on that, but we're going to do it really well. Now, when asked about safety, he downplayed fears of things like self aware AI, but did highlight that agentic AI, you know, systems that can take action on their own, was, quote, the most interesting and consequential safety challenge that OpenAI has faced yet. And when asked about AGI, he said, quote, if you ask 10 OpenAI engineers, you will get 14 different definitions of AGI. Whichever you choose, it is clear that we will go way past that. They are points along an unbelievable exponential curve. Now, Paul, I think you're gonna address this, but like first, what's worth paying attention to here? This interview had some like, nuggets in it, but got kind of funny, kind of awkward because the whole reason we learned that ChatGPT's user base had doubled in just a few weeks is because interviewer Chris Anderson said on stage to Altman, hey, you told me that backstage. And Altman very quickly and annoyingly responded saying, quote, I said that privately.
Paul Raitzer
Yeah, the interview went downhill pretty fast from there. It was honestly one of the more uncomfortable interviews I've ever watched. So Anderson obviously had very specific things he wanted to get at. The vast majority of them were, were highly annoying to Sam, I would say. And, and plus Anderson kept bringing up AI safety people that were on the stage earlier in the show and referencing them. And Sam obviously wasn't like a huge fan of that. So, yeah, it was just really uncomfortable. Like, there's a lot, Sam said a lot. And I think that there was some points that Anderson had every right to maybe push a little bit on for sure. Like the image generation and the impact of memory and the dangers of the models and what they're doing for safety. And like, these are all valid questions. I, I just think that at some point you got to read the room and realize this is going very poorly. And he did not back off at all. Like, it was, it was very uncomfortable. And then Sam started getting very irritated and just like throwing questions back at Anderson at the end, like, well, what do you think? And it was, it was super awkward. But anyway, on memory, a couple things that jumped out to me. Sam said, one day you will talk to Chad over the course of your life. At some point, maybe if you want, it'll be listening to you throughout the day and sort of observing what you're doing and it'll get to know you and it'll become this extension of yourself, this companion thing that just tries to help you be the best you can be. This is not going to Go away. I think that this idea of ever listening, ever watching AI is going to be pushed into society whether we're ready to it or not. Through glasses is primarily I think is the main way. And I just saw this morning that Tim Cook is like all in obsessive now on beating meta at like wearables and glasses and that's like the next frontier for Apple. I My guess is that's what Johnny Ivy and Sam Altman have been working on together is glasses. I think glasses is the most logical vehicle with which AI will be delivered. That's just always on, always listening, always recording, always watching and providing context. So I think all these other wearables are just going to be irrelevant over time. I, I think glasses is the form factor that makes the most sense. So it's kind of weird like honestly thinking about a future where everybody's just got their devices or their glasses. Like watching and listening and remembering everything and being saved into memories and you're what you do, even if you're choosing not to participate is being saved in someone else's memory all the time. It's very sci fi and you know, black mirror esque. But that I'm pretty convinced like that is an inevitable future. Like by the end of the decade probably. Anderson pushed him really hard on the copyright stuff, the intellectual property specifically around like image generation. And Sam needs. I don't know if there's PR people at OpenAI or comms team, but they need real fast to like get their messaging better. Like I will say like it's their messaging on AGI and copyright and intellectual property is so poor. Like as someone who PR communications background, it's like they've never had a single meeting to like figuring out their talking points around AGI and copyright and intellectual property. And it is the core of everything they're going to do to need societal support for what they're going to try and do next. And they can't even vocalize their, their position other than on the image generation stuff. Like they don't think artists have rights basically is pretty much all they're, they're saying. And that is not gonna, that's not gonna fly. Like you can win the court cases maybe, but it's just a very poor approach of like yeah, we need to find a better way to compensate people. Then do it like this. You've been work, you've been at this for like how many years now where you knew this was coming? Do something. Stop saying we need a new model. And then on AGI I just I'm in disbelief at their inability to just say what they think it is. Because every time Sam gets asked this whole like, oh, ask, you know, it's the joke internally ask, you know, 10 open AI researchers, you're going to get 14 different answers. It's like, it's not funny. Like, it's. It is the mission of your organization literally to create AGI and distribute it safely to humanity. What is it? What, what are you actually creating and distributing? I, I don't comprehend that they don't have a consistent answer to this. It's wild. And then the safety stuff. And again, I still get, as I'm saying this, I'm getting worked up. Like, I understand why Chris Anderson was like pushing on these things. I just thought the interview was not handled world great. His views for the future. He said he really believes that society figures out over time with some big mistakes along the way. This is again verbatim how to get technology right. And this is going to happen, basically saying like, these AI advancements, and this is his quote, is like a discovery of fundamental physics that the world now knows about and it's going to be part of our world. And I think this conversation is important, blah, blah. So he's basically saying, like, it's happening with or without us. Like, we now know this is possible and we're going to do it. And then Anderson pushed him on like people leaving the safety team. And Sam basically said, like, yeah, we have different views of what safety is now, in essence. So I think it's an important interview because Anderson asked the hard questions. Sam didn't have good answers to most of the hard questions, which is why I think he kept pushing him. And that itself is sort of maybe illuminative of if that's a word of where we are right now with AI, is that the people building it don't really know what it is that they're building or like what's going to happen when they do. And they just keep pushing forward this idea that we'll figure it out and like we always do, which I think we will. Like, I do think he's right. I think this is. Technology moves, society moves with it. Sometimes it's really uncomfortable. I just really OpenAI has to get a comms team locked in on how to give them talking points because watching these interviews, it's so hard to see their inability to answer what should be fundamental questions to their brand positioning.
Mike Kaput
And I never want to assume malice given how just quickly and chaotic this whole space moves. But it does kind of make you think. I think you alluded to this, like with some of the copyright and the AGI stuff. It's like, really, you have the smartest tools and people on the planet. Nobody has thought of this. Maybe. But at what point is it obfuscating that we don't want to tell you what the real answer is?
Paul Raitzer
Yeah, truly, it's, it's like, I don't know, like it's almost assuming like people are ignorant like that they're, I don't know, it's just lazy. Like, I can't think of another organization where it's been so important to understand what it is they're doing and they just have a complete lack of ability to clarity. It's not funny like that. I think that's what bothers me about it is they laugh it off like, we can't define AGI and yeah, we're stealing stuff, but it's okay that humans have always stole from other humans. Like, no, that's not okay. That can't be the talking point, right?
Mike Kaput
Some other OpenAI news. This past week they are launching a new initiative called the Pioneers program. And the goal of this is twofold. First, help companies evaluate how well AI actually performs in high stakes industries like finance, healthcare and law. And then fine tune models specifically for those use cases. So companies in this program will work directly with OpenAI researchers to design domain specific benchmarks or evals that measure what good performance actually looks looks like in their field. And they'll also get help customizing models through reinforcement fine tuning to basically build expert level AI tailored to narrow tasks in these industries. So Paul, there's a signup form on the page that we link to in the show notes. People can kind of apply to this program. This sure seems like what we've talked about on past podcasts. This need for better ways to evaluate AI's capabilities across real world knowledge work that isn't just related to coding, math or science. And I don't know, it also strikes me as like a sign that OpenAI might be gunning for some high value industry verticals with its models and products.
Paul Raitzer
Yeah. So on episode 141, the Road to AGI, I talked about this like, you know, move past pure IQ tests and into the domain or industry specific tests and it's what they should be doing. Makes total sense. I hope we see more of that. I do think that if I was building a software company for a specific vertical or industry, I would be like really paranoid right now. If I was the VC firm investing in Those companies, I would be asking some really hard questions of the companies that want funding because if OpenAI thinks industry is big enough and they see like billion dollars here, $10 billion there to go build the legal AI or healthcare AI or finance AI, like we already know Microsoft's building the financial analyst AI, like any market they choose to, they have a competitive advantage over the startups that want to do it. So whether they want to do it through a venture fund and they want to fund the building of these and just take an equity stake in some of them, or if they want to just build them themselves, it's going to be pretty powerful. Especially if they're the ones. The proprietary data where they work with people on these evaluations and they know exactly what to build, that's going to be something that the startups don't have if OpenAI chooses not to share it with them. So yeah, I mean this could be the next trillion in market value for them is if they pick off 10 industries and just start building custom solutions for them.
Mike Kaput
One other piece of OpenAI news this week, according to a new report from the Financial Times, OpenAI has drastically cut back the time and resources it devotes to evaluating its most powerful AI systems. Now, what used to take months in terms of evaluation and safety testing now takes days. And insiders say testers were given under a week to assess OpenAI's upcoming O3 model, which is a dramatic shift from the six months of safety checks used for GPT4. The reason, according to the FT, is competitive pressure. As the race heats up with Meta, Google, Xai and others, OpenAI appears to be prioritizing speed to market. Some staff warn that this is a, quote, recipe for disaster, especially as models get more capable and maybe more dangerous. Now, critics also say OpenAI isn't rigorously testing the exact versions it releases and hasn't followed through on promises to fine tune advanced models for biosecurity and other high risk scenarios. OpenAI, on the other hand, claims its new processes are more efficient and still robust. So Paul, I guess we have to accept that OpenAI may indeed be more efficient and just as robust as before in its safety testing. But it sure seems like the incentive is basically just to release as quickly as possible in order to compete, doesn't it?
Paul Raitzer
Yeah. So again, this is one of the issues that Anderson pushed Sam pretty hard on. And you know, again, the basic take was we have evolving views of what safety means, but we use our preparedness framework to evaluate these models and we're not that concerned yet. And we have a track record of iterative deployment, which they think is the safest way to do it, Meaning we're just keep putting things out, see what we learn. And if they reach a level where we don't think they're safe, we won't put them out. And Anders was like, well, how do we know that, basically? And then at one point, Sam, because he got really irritated, I. I think there was might have been like the Ring of Power tweet or something where he was alluding to, like, Sam being like, power hungry and, you know, money hungry. And that was when it was like, you could tell he just was ready to walk off stage. And I don't remember if it was right around there, but he's like, listen, like, all of us care about safety. Well, all of, except maybe one of us, which I can almost guarantee you he's referring to Xai and Groq. He just didn't want to name them directly because earlier in the interview they brought up Elon Musk and saying want nothing to do with it. So, yeah, I think that the concern here is there's no uniform agreement on what safety is, what alignment is. Sundar Pichai, in his interview at Google Next that I was referring to earlier, even said we needed more collaboration on safety, but it's going to be hard to do it because you're going to need to get a lot of egos and a lot of power in a room together to like, talk about these things and agree on it. And it only takes one to push forward and, and do something outside of the bounds. And then the pressures everybody else to decide, are we gonna all go now and do this, or do we, you know, gonna hold our stuff back? So I do think that by the end of 2025, decisions are going to have to be made within the, the Frontier Labs to, to hold back models that have been deemed unsafe internally. Whether or not we hear about that publicly, I don't know. But I'm fairly confident that at some point this year, if it hasn't already happened, these models are going to display some capabilities that hit sort of the red zone of these preparedness frameworks. And they have to hold back and figure out what to do about it. I don't think it's far off at all. And that's just from listening to interviews with like, Dario Amade and others, is these things are moving really fast. And maybe that's part of the GPT5 thing, is, like, maybe it did meet some threshold where they have to hold it back. And put in some more safety standards, I don't know.
Mike Kaput
Meta has started launching models in its Llama 4 family, which are the latest versions of its large language model, but these are arriving a little later than expected and under a bit more pressure than expected. So the headlining models right now are Llama 4 Scout and Llama 4 Maverick. So these are two open source models that can process text, images, video and audio. They use a mixture of experts architecture, activating only select parts of the model per task, which boosts performance while cutting down on compute. Scout fits on a single GPU and supports a 10 million token context window. Maverick, meanwhile, apparently outperforms GPT4O on many benchmarks, though a little more on that in a second and it's optimized for tasks like image understanding, coding and multilingual reasoning. Coming after these models is what they're calling Llama for behemoth, a still training 2 trillion parameter giant model that Meta says will outperform GPT4.5 in STEM tasks. But it is not out yet, though it is serving as a teacher model to boost Scout and Maverick through distillation. Now this launch has a bunch of controversy already around it, because earlier last week Meta began to post that Llama4 Maverick had jumped to the number two spot on the popular LM arena leaderboard. But some users, including prominent AI voice Ethan Malik, started to report that the winning model was different from the version of the model released to users. He actually posted quote, the Llama 4 model that won in LM arena is different than the released version. I have been comparing the answers from arena to the release model. They aren't close. And then the Verge reported that in some fine print, Meta acknowledged that the version of Maverick tested on LM arena isn't the same as what's available to the public. According to Meta's materials, it deployed an experimental chat version of Maverick to the leaderboard that was specifically optimized for conversationality. LM arena then responded, basically saying, yeah, it seems like Meta released a model to our leaderboard that was more customized to human preferences, which play a huge role in the LM arena rankings. So basically a model that others couldn't use, but was designed to rise higher in this specific leaderboard. LM Marine actually said, look, as a result of this, we're updating our leaderboard policies to reinforce our commitment to fair, reproducible evaluation so this confusion doesn't occur in the future. All right, Paul, so it certainly seems like Meta might have ruined their kind of fanfare here by basically gaming a leaderboard. Why would they risk doing this when it was like bound to be discovered?
Paul Raitzer
I don't, I don't know. Plus, they released on a Saturday.
Mike Kaput
Yeah.
Paul Raitzer
Which was weird to start. Like it was. I knew they like they were just maybe getting out ahead of like the Google announcements, but my initial reaction when I was there is like, it must not be very good. Like you're almost like trying to just get it out there and you don't want a bunch of fanfare around it. Yeah, I mean. LM arena tweeted, Meta's interpretation of our policy did not match what we expect from model providers. That was a very PR way.
Mike Kaput
Yeah.
Paul Raitzer
Yeah. I don't know. It wasn't a good week for Meta. They.
Mike Kaput
They're.
Paul Raitzer
And they're getting. They're not having fun in court over the copyright lawsuit either, so. Or. Or the efforts by the government to break them up. It's tough go for Meta at the moment. And yeah, this didn't help things.
Mike Kaput
So kind of related to some of the copyright discussion. There's been some really heated discussion in AI circles this past week related to copyright. It was kicked off, predictably by a controversial post from Jack Dorsey, co founder of Twitter and founder of the financial services company company block, formerly square. On April 11, he posted on X the following quote, delete all IP law. Elon Musk quickly replied, quote, I agree, basically, just like that. Two of tech's influential figures appear to have called for a total teardown of IP protections. It's not exactly clear yet what has prompted this, but this comes as AI companies, including OpenAI, are under legal fire for allegedly scraping copyright material to train their models. As you alluded to, Meta is getting on blast for the same thing. And we've also talked about many times how most, if not all of the model providers have done the same things. Basically, Ed Newton Rex, who we talk about all the time, founder of Fairly Trained, called this a war on creators. A writer, Lincoln Mitchell, put it bluntly, none of Jack Greeland's companies exist by IP law. And Dorsey doubled down arguing that this current copyright system favors gatekeepers over artists. And this comes obviously as all these other copyright things we have talked about have come to pass, where OpenAI drew some heat for its Studio Ghibli kind of viral moments where people were having a lot of fun using its model to generate Studio Ghibli style images, but people also were up in arms about the fact that that style was probably stolen or used without permission. We have controversy about reports that Meta used copyright Books to train its model. And on top of all this, a coalition of major publishers, including the New York Times, Washington Post and Vox Media is calling on the US government to, quote, stop AI theft. So they're launching a sweeping ad campaign this past week that accuses AI companies of, of using their content without permission or payment to train their models. So, Paul, this isn't a new debate. People have been suing over this for a while, but it does seem like it escalated pretty quickly with some tech leaders now feeling comfortable saying we should just delete IP Law.
Paul Raitzer
Yeah, I don't know. I mean, it's certainly a provocative tweet. There's probably a bunch of context behind it. You know, it's like anything else. When you want to argue a point you like, take this extreme position and forget the nuance of the fact that you own copyrights yourself. Your company probably had patents on its initial technology that prevented other people from doing it for years. Like you've made your billions on the top of IP law and now it's convenient to just want to get rid of all IP law because you're a billionaire and it's inconvenient for you to have to, you know, pay people for their creative works. So I don't know, I, it's like I've said many times on the show, I, I have a hard time with extreme positions on anything. I don't care what it is. And so these extreme positions where people pretend like there's no nuance to the conversation just bother me when obviously there's lots of nuance to this conversation. And then, you know, Elon Musk, whatever, like, yeah, it's obviously he's stealing everything he can possibly steal to build Grok, and it's a extreme annoyance to have to potentially face lawsuits over it. So they're just gonna, the way I look at this is like they're gonna just keep taking it. And now like we've seen with image generation from Grok first and then, you know, shortly after from 4,0 image generation with OpenAI, they just don't care anymore. Like they're, they're just all in and they assume it'll work out in the courts somehow or there's going to find some model to pay people back in some class action lawsuits and be done with it. But we are, we are in an accelerated phase of IP theft and that is not going to stop unless the courts somehow stop it. And I just don't see that happening. And I, and I have to be honest, like, you know, we were Using the image generation thing. While I was in Vegas, I was hanging out with a buddy of mine and we were like, I took like a family photo and you can turn into the Simpsons and South park and Anime and Muppets and like Pixar. Like you, you can use all these names now and, and it's fun and it's like really cool to create it. But there is that part of me that's living this, like in this gray area. But like, they're not paying for these. Like, I know you're not allowed to technically do this, they're just doing it. But as a user it's great. Like, it's nice to have the ability to do these things. But as someone who likes, you know, studies the space, I also sometimes like, man, I don't know, I kind of feel guilty about creating these things with these copyrighted images, but.
Mike Kaput
Right.
Paul Raitzer
Yeah, it's a weird space to be in.
Mike Kaput
Our next topic is that Anthropic is offering a new Claude Max plan with two pricing levels. And this is kind of aimed at power users. There's $100 per month for expanded usage or $200 for full access, which basically matches the cost of OpenAI's top tier ChatGPT Pro license. So what you get in return is priority access to new features, including a voice mode that's launching soon, and importantly, significantly higher usage limits, which people have been demanding for a very long time. Anthropic says the demand for this tier has been building for over a year, especially from professionals in finance, media marketing and code heavy fields who rely on CLAUDE to scale up their work. And just a quick reminder here, 200 per month seems like a lot for your average business user, but it seems like there's demand for this. According to some reporting for the information, back in January they actually estimated that ChatGPT Pro licenses, which are the 200 buck a month ones, could be generating as much as $25 million per month for OpenAI. So Paul, should we expect to see more demand and people buying these licenses from Anthropic OpenAI?
Paul Raitzer
I don't know who's paying 200amonth. I mean, maybe it's like developers and stuff, but again, I can only provide the context of like my conversations with big enterprises and, you know, leaders and I don't know anybody using Cloud. Like it's, again, it's. It. I think it's a great product. I still have a subscription to Cloud. I do still test it at times. I just don't. I don't think I have a good context of their market and if they're seeing this as like a competitive product to like the $200 model from OpenAI.
Mike Kaput
Yeah.
Paul Raitzer
And I just, I feel like it's just going to be a really tough battle. Like, I feel like OpenAI has escape velocity with their user base. And again, I don't know, maybe Anthropic is being like super smart about the verticals they're going after, the use cases. I don't know, maybe might be good to like just get somebody from Anthropic on at some point and like hear what it is they're doing, what their, what their market looks like. Because viewing them as a direct competitor to OpenAI is seeming more and more unlikely just given the usage rates at OpenAI. But I could be wrong. But again, I just go back to like, they don't have their own proprietary data, they don't have distribution anywhere. Like two of the critical things that would tell me they're set up to like remain a key competitor aren't there. And I do just continue to wonder about their long term viability as like a major player in this market. But they keep growing. I mean, it's incredible growth and looks like a good company from all things, but whether it competes or not is just hard to say at this point.
Mike Kaput
Next up, we have some more updates on the pretty serious roadblocks that Apple is running into as it tries to build a smarter Siri with advanced AI. So last year Apple promised a smarter AI powered Siri, but according to a new report from the Information Internally, the team appears to not even have been able to agree on the basics. They bounced between models, scrapped a privacy first approach and cycled through leadership changes, all while rivals like OpenAI raced ahead. The chaos led to delays, staff departures, and ultimately the embarrassing admission that the upgraded Siri wouldn't even ship until 2026. Now this article details behind the scenes kind of the fallout as well. Being swift, Apple stripped its AI chief, John Gianandrea of responsibility for Siri, handed control to software head Craig Federighi and Vision Pro exec Mike Rockwell. That team is now pushing to rebuild Siri's future, possibly even opening the door to using open source or third party AI models. It's basically been kind of this hot potato at Apple that's been passed around between teams without a lot of real progress. Now it's back to Federighi, who's known for his execution. But many see it, according to this reporting, as a last chance to bring Apple's AI Assistant up to speed now. Paul, we've talked about this topic at length. This article, though, really seems to pull back the curtain on what's going on at Apple. Can this be salvaged at all, in your opinion?
Paul Raitzer
I don't know. It's Apple. They're incredible company. Obviously they seem to, outside of, you know, recent ups and downs related to tariffs and other things. Like, they, their stock price doesn't seem to really be impacted by their slow move into AI. It's almost like the market is. I've said this before on the podcast. It's almost like it'll be a surprise and an uplift if they ever figure out the AI thing. But they're so strong in product and distribution that it's just kind of like people are kind of right. I was like, okay, the Surrey still stocks and you know, they didn't figure out Vision Pro, really didn't unlock the market we thought they were going to, and yet they just keep humming along as one of the biggest companies in the world. I will say, though, this article was fascinating. Like, the amount of infighting and indecision is hard to fathom for such a meaningful strategic direction, like around AI. So it was definitely the most information I've seen about what happened there and why it happened. And it's hard to, like, look at it and realize, like, they moved so slowly and now you kind of know why for the most part. And then. And again, like, this is just one source. But of all the, the, the publications that Mike and I read and follow to do this podcast every week, the information consistently has stories, often one to two months in advance. And you'll see this in the Wall Street Journal in like two months. Saying, inviting it Apple, like, causes delay. It's like, yeah, the information had that two months ago. You see this all the time. So they're very well sourced and it's a, it tends to be a very factual and credible publication. So I, I do believe that this is probably pretty close to what actually was happening internally. And it's, it's just crazy to look at and realize there's still a ways away. Like, they haven't solved this and like, now we're going to fix it in three months. Like, no, it's 2026 is kind of when they're thinking they're going to write the ship. And that's wild.
Mike Kaput
Our next topic is that the generative AI platform writer who've talked about a bunch on the podcast, has announced something called AI hq, which is designed to help companies build, deploy and manage AI agents at scale. And basically at the core of the AI HQ is something called Agent Builder which is a low code environment where IT and business teams co create agents using visual tools. These agents can then be launched across departments from finance to HR via Rider's new agent library and personalized home dashboards. Apparently, according to Rider, major players like Uber and Franklin Templeton are already using these tools to transform support content, sales pipelines and financial reporting. So Paul, this is just kind of a preliminary product announcement, but it sounds like Ryder, which is a platform we're pretty familiar with, is now all in on agents. Should we be expecting this to kind of become the norm or the direction that all these kind of third party gen AI startups go?
Paul Raitzer
It sure seems that way. That all the SaaS companies are like what used to be apps or templates or just becoming agents. It's like the new terminology and then they, you know, rightfully so they do have new capabilities that are being built into these. So they're not just these deterministic templates. So yeah, I definitely think it's the direction within their announcement. They said there's over a hundred pre built agents. So I just, I clicked through that, I'm just looking at that now and give people an example like so within marketing you've got a product detail page, copy agent, a retail product intelligence brief agent, case study agent, blog post outline agent. So it's a lot of like the tasks that you would do and then they've got it broken into finance. There's like a tax research agent, HR has a job description agent, sales has a earnings call summary agent. So it's, it's task driven, it's industry driven. So it's looking at different, interesting what are the common tasks in this industry and then it's building those and then allowing you to build your own. It's not unlike agent space. We talk about Google where I'm gonna be able to go in and I think theirs is a low code environment. I think a lot of these are gonna be no code to where an average knowledge worker with no coding ability is gonna be able to just go in and build things to, hey, I go through these 10 steps every, you know, Monday morning. Build an agent to do those 10 steps for you. And I, I do think that by this time next year this is all going to be very real. I still think agents are very early and there's probably a little, a little bit of hype and maybe they don't deliver quite what you think they're going to. But I do think by like, end of this year, you know, spring of next year, at anybody in marketing, sales, service, whatever, you're going to be able to just go in and say, hey, I go through these 10 steps, optimize these 10 steps for me and build me an agent that does this for me and then like, have it send me an email every morning. I think that is very much going to be reality that you're going to be able to just automate through prompting with your AI assistant. Just build agents to do repetitive processes and it's going to make work wonderful. I really think, like, anything you can imagine that has a repetitive process, you're going to be able to build an agent to help you do it and to do it way faster than you you did it before.
Mike Kaput
And like some of the examples you just mentioned with Ryder, even though, you know, the company is called Writer, it's definitely moving beyond just kind of marketing content or sales content. Right. Hr, finance. This is touching every single function.
Paul Raitzer
Yep. Yeah, there's a healthcare and life sciences retail and you can tell it's early. Like retail and consumer goods has two agents. Like, yeah, it's all going to get built out and verticalized and yeah.
Mike Kaput
Next up, Mira Murati, former CTO of OpenAI, is back in the spotlight because her new AI startup called Thinking Machines Lab, is reportedly seeking a $2 billion seed round that would make it one of the largest early stage raises in tech history, which is also notable because the company doesn't have a product or revenue and only recently emerged from stealth. It does have a pretty stacked roster of AI talent. Advisors now include Bob McGrew, OpenAI's former head of research, and Alec Radford, one of the key minds behind the original GPT models and OpenAI's Dall E image generation model. Now, Murati says the goal is to build AI that's more customizable, more general and more understandable than what's out there today. They also have on the team ex OpenAI scientists John Schulman and Barrett Zoff as CTO. Paul, this is a pretty staggering seed round, but like not many details which we have also seen with Ilya Startup. Like, what's the bet here that investors are making that makes these numbers make sense?
Paul Raitzer
Yeah, I think I saw in the last couple of days like Ilya just raised another 2 billion at like a $32 billion valuation. And I believe Google and Nvidia were in on that investment round, if I'm not mistaken. Yeah. So I don't Know, like, there's literally tens of billions in value in startups that we don't know what they do. So when you look at this and you look at safe super intelligence, intelligence, familia, we're talking about raising billions with no public knowledge of how they're going to differentiate from OpenAI and others. And I find that fascinating because there has to be something there. There has to be some unique approach to algorithms. There has to be some unique approach to, like, training that's more efficient. Like, it's got to be something it is not. Let's go compete with Google and OpenAI and XAI and Anthropic and spend a billion dollars on a training run. That cannot be what this is because they know they can't keep up. Like, that is. I think that ship has sailed. I think we now know who the frontier model companies will be. Maybe one more shows up or something, but you basically have four or five that can spend the billions to do the massive training runs that, you know, Sam talks about and Google talks about these. These aren't them. I don't, I don't think these are frontier model companies. I think these are something different. And I don't know exactly what it is. There's some, I would make some bets on a couple of things, but I don't know what they are or what their market is. Like. It's so, you know, I was saying earlier, like, what is the anthropics market? Like, what is their total addressable market? What do they look at as, you know, the, the verticals or the industries they're going to go after? And I think the same thing with these is, like, what could they possibly be going after? Because at a $2 billion round, like, if they're like, if you're Ilya, I don't know what the value. What was the valuation on this of the $2 billion?
Mike Kaput
This is 10 billion, I believe.
Paul Raitzer
Okay, so to, So I am not an expert on like investing rounds, but I've, I've spent some time on the, on the topic. So if you're investing at a $10 billion valuation at a seed round, the investors are going to be looking for at least a 10x ray term, like far, probably far greater than that. So they're saying, like, this is a $100 billion company out of the gate, like that the market they see right now is like $100 billion market for this company. Now that's not out of the realm of possibility. If you see Ilias and it's already worth 32 billion. And you see OpenAI just raised 40 billion. Right. So, I mean, there's, there's huge markets out there to be won, but a $10 million, $10 billion seed round is absurd. Like, the growth of that company must be so massive. So it's like, what could you possibly be bringing to market that is that vast in its market potential that you're getting a $10 billion valuation before there's anything to show? It's crazy, right?
Mike Kaput
Yeah. Especially when we're already saying who is paying $200 a month for anthropic. Right. Which is.
Paul Raitzer
Right.
Mike Kaput
Something we understand realistically.
Paul Raitzer
Like as an investor, you have to probably be looking this as a trillion dollar bet. Like you're, you're guessing that this company has the potential to be a trillion dollar company if you're putting that kind of valuation at a seed round with no products, no revenue, nothing.
Mike Kaput
Wow. So, according to a new report from Digiday, agencies, marketing agencies, ad agencies, et cetera, are increasingly adopting deep research tools from OpenAI, Google, Perplexity, among others that we've talked about quite often on the podcast. So just as a reminder, these don't just, you know, do a little research for you. They actually autonomously scan tons of different sources on the web or with data you give it to produce in depth research reports on any topic that at least we found rival or surpass the work of humans. And agencies are apparently starting to take that concept even further, according to Jinjaja. So they're interested integrating deep research capabilities into their proprietary data sets and then using what deep research AI tools produce to then do even more. So one example they cite is Havas, which has integrated deep research tools within their broader data platform. One exec there calls this the long dreamed of quote planning buddy they've always wanted. And then they're taking the insights that deep research produces and turning it into interactive tools like custom GPTs that simulate consumer behavior as kind of digital twins of different audience segments or client types. And they're also using tools to find, synthesize and upload info from a range of external and internal sources to get better insights into their customers, clients and markets. Digiday also mentions one tool outside of the big AI labs that's gaining some traction for its workflows specific to marketing agencies. It's called Waldo, and it claims to be able to automate complex research tasks that agencies do every day. So, Paul, the reason we wanted to mention this is, you know, we've talked a lot about how agencies need to start evolving based on What AI can do today, especially the deep research tools that made this much more obvious that how agencies worked in the past probably cannot be how they work in the future. When I read this, this definitely sounded at least like a step in the right direction.
Paul Raitzer
Yeah, it's cool to see it. I, I'm, I, I do think that there are a lot of agencies that are figuring this out. I, I've talked with some recently, some bigger firms that are doing some pretty cool stuff, like the digital twins idea where you're just creating these segments and you're running simulated campaigns and. Yeah, I just, I think the way this all works out looks so much different a year or two from now. And I go back to, you know, the thing that originally drew me to AI back in 2012 when I was running an agency, was this concept of a marketing intelligence engine where I could put all the data into it and I could actually use an AI. This is way before gen AI, obviously, or in just the machine learning era of AI. Deep learning was just becoming a thing, just being proven that it could. It had the potential to become what it is today. But I had this vision to be able to automate marketing strategy by feeding data in and running simulations and then making predictions about what campaigns would work and generate ROI. So that was what drew me to AI 14 years ago. So to see these sorts of things, it's almost like the early iterations that lead to that intelligence engine I envisioned long ago. I still have not seen someone build that, but I think this is the kind of stuff that starts to get us much closer to where AI is truly infused at the strategic level. And this idea of building Personas and simulations and digital twins, like, that's the stuff that can really start to unlock it, where you can have a million customers represented and then you can run simulations against those million customers to predict the performance of a campaign or tagline, things like that. Hell, OpenAI could do this. They could build a model that runs like, simulations of how people respond to their messaging. Here's 10 ways we're explaining AGI. How would a million people respond to this? Go do that. You could probably find a way to do that. Borrow a few GPUs to run some test messaging against.
Mike Kaput
Yeah, what I also just love about this is how HAVAS is like, not only using the deep research tools apparently, but taking that output and like, the custom GPT for simulated audiences is awesome. But also presumably, it's like, how easy would it be for an agency to run deep research reports on any part of their business or their clients drop a bunch of that into a custom GPT that anyone on the team, even those who aren't as AI savvy, can then start using. Seems like an interesting idea.
Paul Raitzer
Yeah. And imagine like, you know, we just talk about the writer and their AI agents or agents based with Google. Like, imagine an agency or it could be a brand, but whatever. And as a marketer, I just go in and say, hey, I'm trying to plan for a campaign to launch a new product. What GPTs do we have that would help me do this? And like an AI agent you're interfacing with pulls the three custom GPTs that people have built and then recommends which ones will help you and how they'll help you. And here's the rationale, and then asks you, would you like me to start running simulations for you? Would you like me to. That's the future workflow. And I don't think I'm saying far future. I think I'm like 12 months future probably to be able to start doing those kinds of things.
Mike Kaput
All right, Paul. Our last topic today is our recurring segment on listener questions. So each week we take a question from our listeners and our audience and try to answer it to create some more value for everybody. So this week's question, someone asked, how do you filter out the signal from the noise in generative AI, given that the space evolves daily? To which I would also add, given all the hype and craziness that, that we have to follow day in, day out.
Paul Raitzer
Yeah. So, I mean, I'll tell you how I do it and then I'll offer, like a general guidance. So the way I largely stay sane and filter this, I actually have notifications set up on, on X. So I have a highly, highly curated list that I've built over the last 10 years of AI researchers, influencers, authors, entrepreneurs, media, who I trust, who I. Well, who I feel like have some inside knowledge that I don't have. And that by curating that knowledge, I can create a better picture of what's actually happening in AI, that if some announcement is made, I can go to that feed that are purely notifications. This isn't just a list. This is like notifications of, of a list, like, filtered further down. And if something was talked about, something major happened, I can immediately get 10 different perspectives on that announcement. I don't take one person's, like, perspective, like, as the, the truth. I look around and say, okay, what are other people that I follow? What are they saying about this? And I often will put contrarian people into those notifications too. It's like I want to know what the safety people are saying versus the EAC people, the accelerationist. Like, I want both perspectives. So for me to do what we do, where I have to talk about this every, you know, Tuesday, I need to take in as many perspective as possible and be as objective as possible. And the way I do that is largely through highly filtered notifications on Twitter. And then I listen to a ton of podcasts and then I watch a lot, read a lot of research reports and articles, things like that. So basically it starts with filtering people that are your. That are your influencers, who are the people whose voices I trust, and then what are they sharing? What are they talking about? That's how I do it. It. So I've done it for over a decade in AI. If you're just getting started, though, I think you find a couple of voices that you trust. Hopefully we're one of those. Hopefully this podcast each Tuesday is part of your process. I know for some people I talk with, this is their process. Like they don't have time to do everything research this space. So they just listen to this once a week. And that gives them some peace of mind that they're at least aware of the key things that happen. And then I think if you want to go down further, it's like, okay, now you find your newsletters and you find, you know, your people to follow on Twitter or LinkedIn and like, you know, there's standard ways to research it, but that's. That's kind of how I do it, is that I find the trusted voices and then I follow the things that they talk about and the other people that they follow. I. How about you, Mike? How do you keep keep up to date?
Mike Kaput
Yeah, definitely similar. I certainly don't have as built out of a system as you do for the alerts and especially people posting about it. But also I would say what's been really helpful for me aside, trying to stay on top of the news from trusted voices is also taking some time based on what you know or what you're learning to map out. The vision of where you think this is going doesn't have to be like crazy in depth, but like, where do we think in the next 12 to 24 months, your job, your function, your livelihood could be impacted. And that helps me then say, okay, am I looking at something that is noise or something that's interesting but not immediately useful? It gives me a good filter to say, okay, based on where we're going have I made progress towards evolving my career or expertise towards where I think the puck is kind of sliding to? Or am I just spending too much time, you know, reading certain things that are interesting but not going to help me move the needle?
Paul Raitzer
Yeah. And I just, I wouldn't get overwhelmed. I mean, if you're, again, if you're just trying to figure this stuff out, pick one or two things you can ease into it and like figure it all over time. But there's no replacement for experimentation. Like, that's the one thing I would tell people is like you listen to podcasts all day long, take course, whatever you want to do. But until you start using one of these tools, daily chat, Gemini, Claude, whatever you prefer, that's the best way to learn right now is just experiment with things, try them out, figure it out for yourself. What's going on.
Mike Kaput
All right, Paul, that's a pretty packed week in AI. Just a quick housekeeping reminder here. If you haven't checked out the Marketing AI Newsletter marking AI Institute Newsletter, please check that out at marketingai institute.com Newsletter we cover all of today's stories plus all the ones we didn't get to. And Paul, thank you so much for curating and unpacking and demystifying everything for us today.
Paul Raitzer
Absolutely. And quick Show Note no episode on April 22nd I will be on spring break with my family and so there there will be no episode next week. So we will be back. I guess that would be what, April 29th with the next episode, but newsletters will go out in between. I'll probably still Send my Exec AI newsletter from SmartRx AI so you can go to SmartRx AI and subscribe to the newsletter there. That's every Sunday I drop that newsletter with editorial and a preview of what we're going to talk about on the blog or on the podcast. And then Mike authors the newsletter from the institute that has kind of a recap of everything and links to the week. So check both those out. And in the meantime, if anybody else going on spring break, enjoy. I know I'm going to try and unplug and enjoy and we'll be back with you at the end of April. Thanks for listening to the Artificial intelligence show. Visit SmarterX AI to continue on your AI learning journey journey and join more than 100,000 professionals and business leaders who have subscribed to our weekly newsletters, downloaded AI blueprints, attended virtual and in person events, taken online AI courses and earned professional certificates from our AI Academy and engaged in the Marketing AI Institute Slack Community until next time, stay curious and explore AI.
Summary of The Artificial Intelligence Show Episode #144
Release Date: April 15, 2025
In Episode #144 of The Artificial Intelligence Show, hosts Paul Roetzer and Mike Kaput delve into a myriad of pressing AI topics, ranging from ChatGPT's enhanced memory capabilities to significant developments in AI models from major tech players. This comprehensive summary captures the essence of their discussions, enriched with notable quotes and timestamps for reference.
Timestamp: [02:47] - [13:59]
Mike Kaput kicks off the episode by discussing OpenAI’s latest update to ChatGPT: a new memory feature that allows the AI to remember and reference past conversations more seamlessly. Unlike previous iterations where users had to explicitly save memories, the new update enables passive accumulation of contextual insights to enhance future interactions.
Notable Quotes:
Key Insights:
Timestamp: [13:59] - [25:55]
The discussion shifts to Shopify CEO Toby Lutke’s leaked internal memo, which underscores an "AI First" strategy. Lutke mandates that before requesting additional headcount, teams must demonstrate that AI cannot fulfill the required tasks. This move positions AI utilization as a baseline expectation within the company.
Notable Quotes:
Key Insights:
Timestamp: [25:55] - [41:07]
Paul and Mike examine Databox CEO Pete Caputa’s announcement about replacing 80% of customer support and sales development staff with an AI chatbot, resulting in a 40% improvement in results. This move illustrates the tangible impact of AI on operational efficiency.
Notable Quotes:
Key Insights:
Timestamp: [41:07] - [57:07]
The hosts recap major announcements from Google Cloud Next, including the launch of Gemini 2.5 Pro, enhanced AI models, and significant infrastructure upgrades. Paul shares his firsthand experience of the event’s innovative showcases.
Notable Quotes:
Key Insights:
Timestamp: [57:07] - [60:45]
Mike reports on OpenAI CEO Sam Altman’s announcement to release the O3 and O4-mini models before GPT-5, citing the need to manage unprecedented demand and integration challenges.
Notable Quotes:
Key Insights:
Timestamp: [41:07] - [57:07]
Paul and Mike discuss Sam Altman’s recent TED interview, where he revealed ChatGPT’s user base has doubled, now encompassing 10% of the global population. Altman also touched on the challenges of defining and achieving Artificial General Intelligence (AGI).
Notable Quotes:
Key Insights:
Timestamp: [57:07] - [60:45]
The hosts analyze Meta’s release of Llama 4 Scout and Llama 4 Maverick models, which faced criticism for discrepancies between leaderboard performances and publicly available versions. This controversy highlights challenges in AI model transparency and fairness in evaluations.
Notable Quotes:
Key Insights:
Timestamp: [53:00] - [60:45]
OpenAI introduces the Pioneers program aimed at assisting companies in high-stakes industries—such as finance, healthcare, and law—in evaluating and customizing AI models to meet domain-specific needs.
Notable Quotes:
Key Insights:
Timestamp: [53:00] - [57:07]
Recent reports indicate OpenAI has expedited its safety and evaluation processes for new models, reducing evaluation time from months to days to meet market demands. Critics express concern over potential safety compromises.
Notable Quotes:
Key Insights:
Timestamp: [60:45] - [72:26]
Paul and Mike discuss Apple’s delayed efforts to enhance Siri with advanced AI, citing internal conflicts, leadership changes, and strategic indecision as major hurdles. The revamped Siri is now scheduled for a 2026 release.
Notable Quotes:
Key Insights:
Timestamp: [72:26] - [85:41]
The episode highlights Writer’s new AI HQ platform, designed to help companies build, deploy, and manage AI agents at scale. Features include a low-code Agent Builder and an extensive agent library catering to various departments.
Notable Quotes:
Key Insights:
Timestamp: [76:07] - [80:29]
Anthropic unveils the Claude Max plan, offering two pricing tiers aimed at power users with expanded AI usage and new features like voice mode. This move aligns Anthropic’s offerings with premium models such as OpenAI’s ChatGPT Pro.
Notable Quotes:
Key Insights:
Timestamp: [76:07] - [80:29]
Former OpenAI CTO Mira Murati’s new startup, Thinking Machines Lab, seeks a staggering $2 billion seed round. Despite having no product or revenue, the company boasts an impressive roster of AI talent.
Notable Quotes:
Key Insights:
Timestamp: [80:29] - [85:41]
The hosts explore how marketing and ad agencies are integrating deep AI research tools from providers like OpenAI, Google, and Perplexity to enhance their operations. Firms like Havas are leveraging AI to create interactive tools and simulate consumer behavior through digital twins.
Notable Quotes:
Key Insights:
Timestamp: [85:41] - [90:31]
Addressing a listener’s query on managing the overwhelming influx of AI information, Paul and Mike share their strategies for filtering valuable insights from the constant stream of updates and hype.
Notable Quotes:
Key Insights:
Episode #144 of The Artificial Intelligence Show offers a deep dive into the latest AI advancements, strategic corporate shifts towards AI integration, and the evolving landscape of AI development and deployment. With insightful discussions on the implications of enhanced AI memory, industry-specific AI solutions, and the competitive maneuvers of leading tech giants, Paul and Mike provide listeners with a nuanced understanding of the current AI ecosystem. Their emphasis on AI literacy, strategic adoption, and the importance of filtering information amidst rapid advancements underscores the critical considerations for businesses and professionals navigating the AI-driven future.
For those looking to stay updated, the episode also highlights the importance of curated information sources and ongoing experimentation with AI tools to effectively harness their potential.
This summary was crafted based on the transcript provided and aims to encapsulate the key discussions, insights, and conclusions from Episode #144 of The Artificial Intelligence Show.