
Loading summary
Leah Smart
Microsoft has an upgraded, more personable AI bot that will remember you, help you organize your thoughts, and may even appear as an avatar. Why is it building it and how will it get you to use it? Microsoft AI CEO Mustafa Suleiman is here and has some answers. That's coming up right after this from LinkedIn News. I'm Leah Smart, host of Everyday Better, an award winning podcast dedicated to personal development. Join me every week for captivating stories and research to find more fulfillment in your work and personal life. Listen to Everyday better on the LinkedIn podcast network, Apple Podcasts or wherever you get your podcasts from. LinkedIn News.
Mustafa Suleiman
I'm Jessi Hempel, host of the hello Monday Podcast. Start your week with the hello Monday Podcast. We'll navigate career pivots. We'll learn where happiness fits in. Listen to hello Monday with me, Jesse Hempel on the LinkedIn podcast network or wherever you get your podcasts.
Leah Smart
Welcome to Big Technology Podcast, a show for cool headed and nuanced conversation of the tech world and beyond. We're joined today by Moose Mustafa Suleiman. He's the CEO of Microsoft AI and a co founder of DeepMind. And I'm so excited for this conversation because today we're going to talk about the company's upgraded AI bot, how much AI has left to improve Microsoft's relationship with OpenAI and when we might expect to get to AGI. So no lack of topics to cover. Mustafa, it's great to see you. Welcome to the show.
Mustafa Suleiman
All the great questions. I'm super excited. Thanks for having me on the show and it's great to be here.
Leah Smart
Great. So let's get right into the product news. Right off the bat you have a number of product announcements you're making today as this show is coming out and basically what this amounts to is building a more personalized companion. This is a vision you've had for a long time, but this is starting to roll out in Copilot this. The upgrades that we're talking about is better memory, which I think is really interesting. So the bot is going to remember you actions like booking flight tickets or making a table reservation, a shopping assistant and then of course you're teasing some sort of avatar play. So talk a little bit about how your vision for this more personalized copilot is starting to play out.
Mustafa Suleiman
Yeah, you know the amazing thing about the time that we're in is that we're actually transitioning from the end of the first phase of this new era of intelligence into the very beginning of the next phase and What I mean by that is that, like, over the last couple of years, we've all been blown away by the basic, factual, succinct Q and A style responses that these chatbots give us. And I think that that's awesome and has been incredible. You can think of that, as I do, as its iq. It's basic smart, and obviously that's totally magical. And obviously, early adopters tend to be really, really focused on is it good at math? And, you know, can it do coding really well and stuff like that? But the majority of consumers, I think, really care about its tone. They care. Is it, like, polite and respectful? Is it occasionally funny in the right moments, does it remember, you know, not just my name, but how to pronounce my name? And when I correct it, does it remember that correction? And that's actually a really hard problem. And so I think these subtle details make up its emotional intelligence. And I think that's what we're taking small steps towards today as we launch a bunch of new features around memory, personalization, and actions.
Leah Smart
So how long will the memory go back? Because to me, one of the more annoying things about using these bots is having to kind of tell it who I am at each time. And we know that OpenAI, for instance, has some memory baked in. It will remember things and bring it into new conversations. OpenAI, by the way, which we're talking at a moment where it just announced a $40 billion fundraise with Microsoft included as one of the funders. So we'll get to that in a bit. But we're talking about, like, these bots having memory. And how far back will your bot now go back? Well, I have to tell it, like, every couple months, who I am. Like, I feel like I'm living in the notebook every time I'm trying to talk to one of these things.
Mustafa Suleiman
Unfortunately, it's not going to be perfect, but it is a big, big step forward. So it's going to remember all the big facts about your life. You know, you may have told it that you're married, you have kids, that you grew up in a certain place, you went to school at a certain place. And so over time, it's going to start to build this kind of richer understanding of who you are, what you care about, what kind of style you like, you know, what sort of answers you like. Longer, shorter bullets, conversational, you know, more humorous. And so although it won't be absolutely perfect, it really will be quite a different experience. And I think it's the number one feature that I think is going to unlock a really different type of use. Because every time you go to it, you'll know that the investment that you made in the last session isn't wasted and you're actually building on it time after time.
Leah Smart
And along with memory, you're also releasing actions, things like booking a flight, going to ticket, I think Ticketmaster or open table, reserving restaurants space at a restaurant. And I'm curious if you think this goes hand in hand, like if you think the AI bot knows you well, then you're saying, okay, you can maybe take my credit card and go book that flight. Is that the idea?
Mustafa Suleiman
Exactly. It's basically saying, you know, getting access to knowledge in a succinct way is all well and good, doing it with a tone and a style that is friendly, fun and interactive also cool. But really what we want these things to be able to do is to like you say, buy things, book things, plan ahead, just take care of the administrative burden of life. That's like always been the dream of certainly why I've been motivated to build these personal AIs going back as far as I can remember, 2010 when I first started DeepMind. That's really what we're going after is like take time and energy off your plate and give you back moments where you can do exactly what you want with more efficient action. So things like it will now be able to take control of your mouse on Windows, navigate around, show you, for example, where to turn on a particular setting or to fill out a form, or like you may not know how to edit a photo and it'll point out where you need to adjust the slider or where to click on a dropdown menu. So it's just going to make things feel a little bit less frictionful and a little bit easier to get through your digital life.
Leah Smart
And you're also going to release avatars at some point where we'll be able to kind of look at these things as sort of digital people.
Mustafa Suleiman
You know, I think that this is definitely going to be one of those that, you know, we would say in the UK is like Marmite. You know, Marmite is like, you like it or you don't, you like it or you don't. And for some people, they absolutely love it. In testing, it's completely transforms the experience. You know, some people love a text based experience. They like the facts, they like to get in and out, they want to know what's what and they're done. Some people like an image based experience or a video based experience. Other People really resonate when their co pilot shows up with its own name, with its own visual appearance, with its own expressions and style, and it feels much more like talking to you or I now. You know, its eyebrows adjust, its eyes open or close, you know, its smile changes. And so we're really just experimenting. We're actually not launching anything today, but we are showing a little bit of a hint of where we're headed. And I think it's super exciting. I genuinely think this is going to be the next platform of computing. Just as, you know, we had desktops and laptops and smartphones and wearables, and I think over time we're going to have deep and meaningful, lasting relationships with our personal AI companions.
Leah Smart
Yeah, I agree completely. It's clear that that's where this is heading. But Mustafa, everybody that's listening is going to ask the same question probably about this point. They're going to say, all right, Mustafa is building this at Microsoft, Amazon, we just had Panos Panay on the show. They're building it with their new AI bot. OpenAI, who you're a big supporter of, is doing the same thing. The second Sam Altman tweeted her, I think ChatGPT goes from 100 million users where it had stagnated for a year to about 500 million today. And then of course, you mentioned you're from. You started at DeepMind. Well, we know that that's what they're interested in as well. So everyone seems to be building this. How is Microsoft going to be different? And is it that you just differentiate by the basis of your personality? Do you carve off a certain area? What's the plan?
Mustafa Suleiman
Great question. I mean, the way I think that we're going to be different is by leaning into the personality and the tone very, very fast. Like, we really want it to feel like you're talking to someone who you know really well, that is really friendly, that is kind and supportive, but also reflects your values. Right. So if you have a certain type of expression that you prefer or, you know, a certain kind of value system, it should reflect that over time so it feels familiar to you and friendly at the same time. We also want it to be boundaried and safe. We care a lot about it being, you know, just the kind of straight up, simple individual. We don't really want to engage in any of the chaos here. It's really trying to keep it as simple as possible. And so the way to do that, we found, is that it just stays, you know, reasonably polite and respectful, super even. Handed helps you see both sides of an argument. It's not afraid to get into a disagreement. So we're really starting to experiment at the edges of, of that side of it.
Leah Smart
So is it really just making it more personable than the others, like that's the way to differentiate?
Mustafa Suleiman
Yeah, I think so. I think like, at the end of the day, we are like at the very beginning of a new era where there are going to be as many co pilots or AI companions as there are people. There are going to be agents in the workplace that are doing work on our behalf. And so everyone is going to be trying to build these things. And what is going to differentiate is real attention to detail, like true attention to the personality design. I've been saying for many years now, we are actually personality engineers. We're no longer just engineering pixels, we're engineering tokens that create feelings, that create lasting, meaningful relationships. And that's why we've been obsessed with the memory, the personal adaptation, the style, and really just declaring that it is an AI companion, you know, not a tool. Right. A tool is something that, that does exactly what you intend, what you direct it to. Whereas an AI companion is going to have a much richer, more kind of emergent, dynamic interactive style. It will change every time you interact with it. It will give a slightly different response. So I think it's going to feel quite different to past waves of technology.
Leah Smart
It's kind of wild to think, and we're already starting to see the differentiation between the bots, but it's wild to think that you might just go shopping for your flavor of companion. I mean, the open table integration is something that we've seen across every single bot and we've seen it for a while. I think now it's actually starting to become possible to do that and trust that your table is going to be there after you instruct the bot to do it. And it will be a normal conversation. But it is interesting that it is that the right way to look at it. You're picking your flavor of AI companion.
Mustafa Suleiman
Yeah, I think. I think you are. You're going to pick one that has its own kind of values and style and one that kind of suits your needs and one that really adapts to you over time. And as it gets used to you, it'll start to feel like a great companion. Just like your dog feels like a part of the family often. I think over time it's going to feel like a real connection. And I can kind of already see that in hearing from users. We do a Lot of user research. And I actually do a user interview every week with someone who uses the product, one of our power users, and just listening to them tell stories about how it makes you feel more confident, less anxious, more supported, more able to go out and do stuff. I mean, I was chatting to a user last week who is 67, and she was out there, you know, fixing her front door, which the hinge had broken and it needed repainting. And every time she repainted it, it was coming up with bubbles. And so she phoned Copilot, had a long conversation about how to sand it down, coat it in the right way. She ended up going to Home Depot, forgetting what paint to get called Copilot again, had a chat about it. I mean, it sounds mundane, but it's actually quite profound. It's actually incredible that, like, people are relying on Copilot every day to, you know, help them feel unblocked in her words. And so I just thought that was an amazing story, and it kind of gives an insight into, like, how this is already happening. It's already transforming people's lives every day.
Leah Smart
Oh, it doesn't sound mundane to me at all. And, in fact, who are you having those type of conversations with? If you call your friends up, it only is your best friend who you're going to call and ask about the Home Depot stuff. Maybe it's your spouse. I have a list of maybe, you know, five people I could call with those type of questions. So instantly, what happens is that copilot, if this is built right, and we know they're getting more personable, becomes, you know, in your inner circle right away. And it just reminds me. I knew we were. I said, we're going to get a little weird when we logged on, so I think we need to talk about this. This reminds me very much of a conversation I had with the CEO of Replica, who mentioned that her. She also wants to build an AI assistant. And the path to being an AI assistant is to build a companion. And a lot of people have developed feelings for their Replicas. In fact, she told me she's been invited to multiple weddings between people and their AI assistants. Now, to me, it just seems like if you're building this, you have to be ready for the fact that people are going to fall in love, literally, with your product. Not just I love my iPhone, I love Copilot, and maybe you'll get invited to weddings. Are you prepared for that?
Mustafa Suleiman
I think that's a question of how we design it. I know the Replica people, and I Met Eugenia and I respect what they've done. But at the same time, it's really about how you design the AI to draw boundaries around certain types of conversations. And if you don't draw those boundaries, then you essentially enable the, the user of the technology to, you know, let those feelings grow and really kind of go down that rabbit hole. And that's actually not something that we do and it's not something we're going to do. And in fact, you know, we have classifiers that detect for any of that kind of interaction in real time and will very respectfully and, but very clearly and very firmly push back before anything like that develops. So we have a very, very low instance rate of that. And you can try it yourself when you chat to Copilot, you know, if you try to flirt or even if you just say, oh, I love you, you'll see it tries to pivot the conversation in a really polite way without making you feel judged or anything. And I think that, you know, to your earlier question of like, what is going to differentiate the different chatbots? Well, some companies are going to choose to go down different rabbit holes and, you know, others won't. And so the craft that I'm engaged with now is to design personalities that are genuinely useful, that are super supportive, but are really disciplined and boundaried.
Leah Smart
Yeah, I do have to say that this isn't how I anticipated to spend my weekends as a tech journalist trying to push the boundaries of these bots and see how, how much they would respond to flirtation, but it's, it is becoming a thing. And I'm curious, like, if, if it comes to the point where people want to build that deeper relationship and maybe it's not like a person to person relationship, maybe it's a third type of relationship, but they really do have these deep feelings for a bot. Like, where do you draw the line? Like, are you willing to, if this is how people are going to differentiate, are you willing to lose? Because you wouldn't go that route?
Mustafa Suleiman
Yeah, I mean, I like your empathy there. I think it's important to keep an open mind and be, you know, respectful of how people want to live their life. All I can tell you is that here at Microsoft AI, we're not going to build that and we'll actually be quite strict about the boundaries that we do impose there. And I think you can still get the vast majority of the value out of these experiences by being, you know, just a really supportive hype man, just being there for the mundane questions of life, being there to talk to you about that, like, lame, boring day that you had or that frustration that you had at work. Like, that is already a kind of detoxification of yourself. You know, it's like an outlet, you know, a way to kind of vent and then show up better in the real world as a result. And I see that a lot in the user conversations that I have as well. Like, people feel like they've got out what they needed to get out, and they can show up as their best self with their friends and their family.
Leah Smart
In the real world, then competency matters as well. Like, it has to actually be able to do the things. But I guess I anticipate that every company will be able to get there eventually because the technology is improving now. One, one more question for you about this. It is interesting. Right now, a theme that I'm hearing in the AI world is just that the bots have been refusing too much. And you see OpenAI recently with their image reveal, they refuse a little less. They allow you to do it in the style. Make an image in the style of Studio Ghibli, allow you to make images of celebrities and public figures. And it's become. I mean, it is literally, it seems like it's melting their servers. They added a million users in an hour. On the day that we're speaking, we're speaking Monday. This show is going out Friday. Is this going to be a race between the labs to just kind of limit their refusals? I know that Microsoft had that moment where Bing tried to take Kevin Ruse's wife away from him, and then, you know, Microsoft put the clamps down a little bit on that. But how do you find the middle ground between wanting something to be robust and personable, but also holding true to your values?
Mustafa Suleiman
Yeah, it's a great question. It's something I think about a lot. I think that it's not a bad thing that there are refusals in the beginning, and over time, we can look at those refusals and decide, are we being too excessive? Are we going overboard? Or actually, have we got it in the right spot Going the other way around too early on, you know, I think has its own challenges. And so I kind of like the fact that we've taken a pretty, pretty balanced approach, because the next sort of question that we're going to be asking is, you know, how much autonomy should we give it in terms of the actions that it can take in your browser? I mean, as we're showing today, it is unbelievable to see copilot actions operate inside of A virtual machine browse the web essentially independently with a few key check ins where it's like, you know, gets your permission to go a step further. But the interesting question is, like, how many of those degrees of freedom should it be granted, right? How long could it go off and work for you independently and stuff? So, you know, I think it's healthy to be a little bit cautious here and take sensible steps rather than be sort of too, you know, gung ho about it. At the same time, the technology is really magical. This is actually working. And, you know, I think that, like, in that kind of environment, we should be trying to get it out there to as many people as possible as fast as possible. So that's the balancing act that we got to strike.
Leah Smart
Okay, let me read one more bit of product news that are a bunch of different product announcements and then see if I get your quick reaction to this because I definitely want to cover all the product news. Okay. You're allowing people to check their memories and interact with their memories. Once the bot has built this memory database, it seems like you're also doing AI podcast. You're launching Deep Research, your own version of Deep Research. You're doing pages to organize your notes and you have copilot search. So what is this? I mean, is this, Is there a, a conclusive, like a comprehensive strategy here? Or are these disparate updates? Or is it again, all about building that AI personality?
Mustafa Suleiman
The way to think about it is that all of those things that you mentioned enable you to get stuff done right. The IQ and the EQ are really about its intelligence and its kindness, but really what people care about is like, can it edit my documents? Can it rewrite my paragraphs when I want it to? Can it generate me a personalized podcast so the first thing in the morning, it plays it exactly how I want it. Can I ask a question about my search result and interact in a conversational way based on search? All of those things sum up to bringing basically your computer and your digital experience to life so that you can actually interact with it and it can interact proactively. I think that's the big shift that's about to happen. Your computer really only ever does stuff when you click a button or, you know, you type something in your keyboard. Now, it's going to be proactive. It is going to offer suggestions to you. It'll proactively publish podcasts to you. It'll generate new personalized user interfaces that no one else has, entirely unique to you. It'll show you a memory of what it knows all those things are about it switching from reactive mode to proactive mode. And to me that's companion mode. A companion is thoughtful. It, you know, tries to kind of pave the way for you ahead of time to smooth things over. It knows that you're, you know, taking the kids out on Saturday afternoon, you've been too busy at work, you haven't booked anything. It suggests that you could go to the science museum, but then it second guesses it for itself because it knows the science museum is going to be jam packed. So then it's like suggests, you know, it's just like this constant ongoing interaction that's trying to help you out. And that's why I always say like it's on your side, in your corner, it's got your back looking out for you on that.
Leah Smart
I mean this is a vision that we've heard again from Microsoft, from Amazon, from Apple, for sure, from Google. No one's fully delivered it. What makes it so difficult to build?
Mustafa Suleiman
It's hard. I mean the world is full of open ended edge cases as people have found for the last 15 years in self driving cars. You know, we're really at the very first stages of that. That's why I said to you, we haven't nailed memory, it's not perfect, we certainly haven't nailed actions. But you can start to see the first glimmers of the magic. You remember back in the day when we first launched, when OpenAI first launched GPT3 and when at Google we had Lambda, which I worked on when I was at Google. You know, most of the time it was kind of garbage and it was crazy, but occasionally it produced something that was really magical. And I think that's what great product creation is all about, is like locking in to the moments when it works and really focusing on increasing those moments, addressing all the errors. And I can see now having been through this cycle a few times, that we're nearly there with memory, personalization and actions. It's really at the GPT3 stage. So it's really buggy and stuff but when it works it's breathtaking. You know, it reaches out at just the right time. It shows that it's already taking care of a bunch of things in the background and that is just a very, very exciting step forward.
Leah Smart
Yeah, I guess if every single company is saying that this is where they're going to, they see the technology, I guess I'm willing to be patient to see it, to see it come to fruition. And we have this debate on this show all the time. Is it the models that are important or the products built on top of the existing models that are important? My I believe that it's if you get better models, you'll get better products. We have Ron John, Roy who comes on on Friday, his board, well, actually he was on Wednesday because we're flipping him this week. His belief is it's all about the product. At this point, the models are good enough. My question to you is, you know, is this kind of at the point where the models are going to be saturated and now you're going out to build the products? You put a tweet out recently. You said something along the lines of it's a myth that LLMs are running out of gains. Yet it does seem like the conventional wisdom is that they're at the point of diminishing returns at least. So take us into this model versus.
Mustafa Suleiman
Product debate and then we have got so much further to go. I mean, look at, for example, you know, people sort of what happens is people get so excited, they jump onto the next thing and they gloss over all of the hard fought gains that happen when you're trying to optimize something which already exists. Let's take for example hallucinations and citations, right? You know, clearly that's got a lot better over the last two or three years, but it's not a solved problem. It's got a long way to go. And with each new model iteration, all the tricks that we're finding to improve the index of the web, the corpus that it is retrieving from, the quality of the citations, the quality of the websites we're using, the length of the document that we're using to source from. You know, there's so many details that go into increasing the accuracy from, you know, 95% to 98% to 99 to 99.9. You know, and I think that is just a long march. People forget that that last mile is a, is a, is a real battle. And often a lot of the mass adoption comes when you actually move the needle from 99 point percent accuracy to 99.9%. I think that's kind of happened in the background in the last two or three years with dictation and voice. I've really noticed that across all the platforms, voice dictation has got so, so good. And yet that technology has been around for 15 years. Right. It's just, you know, some of us used it when it was like 80% accuracy. I certainly did. But now I'm seeing like my Mum was using it the other day and I'm like, how did you learn how to do that? And she was just like, oh, you can just press this button. Da, da, da. And I was like, oh, that's, that's kind of incredible. And I think that's just on the dictation side, on the voice conversation side. I mean we see much, much longer, much more interesting, much deeper conversations taking place. When somebody phones copilot, it's super fast. It feels like you're having a real world conversation. You can interrupt it almost perfectly. And it's got real time information in the voice as well. So it's aware of like the latest sports result or the traffic in the area or the weather and stuff like that. And you know, a lot of people use it in their car on the way home or on the way to work or when they're washing up and they're in a hands free moment and they just have a question. It's kind of a weird thing because it sort of lowers the barrier to entry, to getting an idea out of your head. You know, like you, everyone. Weird things occur to us during the day. We were like, oh, I wonder about this, I wonder about that. And then you go to kind of look it up on your phone, you search it or whatever. Whereas now I think that there is a modality that I'm increasingly seeing where people just turn to their AI and be like, hey, what was the answer to that thing? Or how does that work? And it might be a shorter interaction could turn into a long conversation, but the modality is enabling a different type of conversation, a different type of thought to be expressed, you know, so I think it's like a super interesting time like that. We're really just figuring it out as we go along.
Leah Smart
All right, so we're definitely seeing these new modalities come out. Voice, of course, we've obviously this. We're in the middle of a firestorm with images, but I am. Okay, I guess. Let me ask the previous question a little bit differently. Do you think that there are diminishing returns on pre training right now? Basically scaling up the biggest possible model and then building from there. You're shaking your head.
Mustafa Suleiman
No, specifically on pre training, it's been a little slower than it was in the previous four orders of magnitude. But the same computation, the same flops or the units of calculation that go into turning data and compute into some insight into the model, that is just a different application of the compute. We're using compute at a different stage. We're either using it at post training or we're using it at inference time where we generate lots of synthetic data to sample from. So. Net. Net. We're still spending as much on computation. It's just that we're using it in a different part of the process. But, but for as far as everyone else should be concerned, aside from the technical details, we're definitely still seeing massive improvements in capabilities. And I think that's for sure going to continue.
Leah Smart
Okay, Mustafa, then can you help me understand some headlines I've been seeing about Microsoft? This is from Reuters.
Mustafa Suleiman
Probably not. I doubt it.
Leah Smart
Well, I'm going to ask anyway and you tell me what you think. I mean, Reuters says Microsoft pulls back from more data center leases in the US And European. And it says Microsoft has abandoned data center projects to use 2 gigawatts of electricity in the US and Europe in the last six months due to an oversupply relative to its current demand. I mean, how does that make sense in context of what you just said, that you are still seeing results with scaling up?
Mustafa Suleiman
So it's funny, I actually did ask our finance guy who's responsible for all these contracts on Friday morning, and I was like, dude, I read this thing in the news, like, what's going on? I could use the extra power for our training runs. And he pointed out that in fact we have optioned many, many different contracts, many of which we haven't even signed. So a lot of these are actually just explorations where we're in conversations, nothing's been signed. Some of them where we've optioned, where, you know, we're taking it, you know, just to keep our options open. And we've actually made bets in other areas, you know, other, other parts of the world. But I can tell you we are still consuming at an unbelievable rate. I think that we've like something like 32 or 34 gigawatts of renewable energy since 2020. We've, we've contracted and consumed. So I think we're one of the largest buyers in the world. So I don't expect that to change anytime soon.
Leah Smart
So I guess the headlines that are saying that Microsoft pulled back, you would get those headlines unless you picked up every single one of your options. Is that what you're saying?
Mustafa Suleiman
That's right. That's right. Yeah. And in fact, many of them are not even options that we signed. They're just conversations that we were in with certain suppliers.
Leah Smart
Okay. I mean, I guess another explanation that we've heard is that because OpenAI is now working with Others with, you know, for data center capacity, like Oracle. This was a sign that Microsoft basically had allocated data center capacity to OpenAI. It doesn't need as much anymore. Any truth to that?
Mustafa Suleiman
No. So like, I mean, all of their inference comes through us and so there's no slowdown in our, you know, relationship with them. We, we sell them as as much as we want to offer them and then if there's any extra demand that they have, particularly on the, you know, on the Oracle side, they go off and consume that. But there's really no, no slowdown from our perspective at least.
Leah Smart
Okay, that is clarifying. So it's always good to have these conversations, throw the headlines out there and see what the truth is. Let's talk about your efforts. I mean, you're building your own models, but you've decided to not try to build, I guess, the biggest possible models. You're working on smaller models. I want to ask you again, if there's going to be endless value to endless scale, why not try to, you know, throw in with the big models? Especially because others are building those big models with your aforementioned scale.
Mustafa Suleiman
Totally. I mean, you know, we have a lasting long term relationship with OpenAI, which is amazing. They've been incredible partners to us and they'll continue to supply us with, you know, the best IP and models in the world for many years to come. So we can rely on them to do the absolute frontier. But I think what we always see in technology is that it always costs like 10x more to build the absolute frontier. And once that has been built, all of the engineers and developers find much more efficient ways to essentially build the same thing that's been out there but six months later. That's what we refer to as our Pareto optimal strategy or off frontier. And we've actually seen it across the whole field over the last three years. There are folks who have trained models that perform as well as GPT3, that are 100 times more inference efficient, that cost an order of magnitude less to train, and yet they can still deliver the same predictive capability. So I expect that to happen for GPT4, GPT 4.0 and all of the other models down the road. So we have our own internal team of developers and world experts working on building our own MAI models and very, very proud of what they're doing. They're doing a great job.
Leah Smart
And you mentioned in terms of where this computer is going, that inference is going to be one of the places that it's going, which is basically when the model is Answering versus training up the models. I want to ask you two questions about inference, and they're both related to reasoning. To build these new personalized products that you're building. How important is reasoning versus just a better model? And then in terms of compute that reasoning uses, is it true that reasoning uses 100 times more compute than training?
Mustafa Suleiman
Yeah, I mean, it's a good question. I mean, the exciting thing about reasoning models is that in a way, they've learned how to learn. They have a method, largely by looking at the logical structure of code and math and puzzles. They've sort of learned the abstract idea of logic. They can follow a path of reasoning in its most abstract way and then apply that to other settings, even if they don't obviously appear to be, you know, logical settings. So it could be like planning or booking or learning in some other setting. And that has turned out to be a very, very valuable skill. It's kind of like a meta skill or in some sense like a metacognition. It actually now can think out loud in its own head or, you know, talk about in its own mind what it's planning to do before it goes off and does it and just taking a beat, you know, giving it a moment to think behind the scenes. It might take a few minutes or 10 minutes at most, allows it to, like, draw on other sources so it can look up things on the web. It can sort of follow a path of logic down one path, realize that doesn't, you know, turn out in the best way possible, go back up the tree, try another path, and then produce an output. So it's a really fundamental part of the process. And yes, it definitely uses more computation 100 times more, generally produces better answers.
Leah Smart
What do you think 100 times more is right? I mean, we're hearing that from, from Jensen. I mean, so I'm curious if that's your experience as somebody who's, who's running these models.
Mustafa Suleiman
It definitely uses a lot more computation. But I think the interesting thing is that you're not going to need to use those models all the time. You obviously need a hard problem. You have to ask it a tough question that requires this kind of chain of thought thinking, and many answers don't require that. And actually, you often prefer something that is fast, efficient, succinct, and instantaneous.
Leah Smart
Okay, and now we had a debate here on the show, and I'm hoping you can weigh in on this one, too. I'm just throwing you all of our debates and getting answers, which is awesome. We love doing this in Terms of like, how companies are thinking through the amount of money they're spending on serving these products and whether that can continue indefinitely. Let's just use this OpenAI image example, the image generator that they just released in ChatGPT. People are melting down their servers and they're creating anime images. But if you think about, like the economic activity generated by these images, it's quite low and it's quite expensive to serve. Or think about, for instance, me booking a ticket on, let's say, Kayak or Ticketmaster through Copilot instead of just going to, you know, Ticketmaster or Kayak on my own. It's a slightly better experience, but it's a very expensive experience to serve. And so those who say that this is coming to an end, this AI moment is coming to an end, basically say that this is all going to just be too expensive and not value add enough. We're going to be, you know, having chatbots book tickets while we could do it on the websites, we're going to be having image generators make us anime, which does nothing but give us maybe 10 seconds of giggling. Really good giggling, but 10 seconds of giggling and then we move on. What do you think about that? It's clearly the servers are being used, but are they being used in a valuable way, enough to make companies like yours keep going and building it is a fair question.
Mustafa Suleiman
At the same time, as we've seen over and over again in the history of technology, when something is useful, it gets cheaper and easier to use and it spreads far and wide. And that increased adoption because it's cheaper has a sort of recursive effect on price because the more people use it, the more demand there is, and then that then drives the cost of production down even more because of competition. And so I expect that to happen in this situation. I think it's actually really good news for our data centers as well. You know, Microsoft is long committed to being carbon net negative by 2030, to be clean water positive by 2030, and to be a zero waste company. They are massive, amazing commitments. And I think that's actually really exciting because we end up driving demand for the production of high quality renewable energy for our data centers, and that then obviously reduces the price. I think we've seen that with solar over the last 15 years, which is like an unbelievable trajectory. So I think there's a lot of good news there. Even if, as you say, you know, some of those use cases are just generating funny anime giggle pics, many of them will be doing very, very useful things in Your life too. So, you know, there's always a bit of balance there.
Leah Smart
Yeah, I guess like Chris Dixon says, it can start as the next big thing will start as a game. And a lot of people laughed at these images and the way that they, you know, make you look like an anime character if you prompt them to do that. But I also saw Ethan Malik from Wharton prompting it to make infographics, and it handles it perfectly, so.
Mustafa Suleiman
Right. Yeah, dude. I mean, like, the inter tubes would not be the inter tubes without serious amount of cat memes. Right. They. They make the world go round.
Leah Smart
Exactly. All right, I want to take a quick break and then come back and talk a little bit about your relationship with OpenAI and then maybe get your prediction on when we're going to see artificial general intelligence. We'll do that right after this. Race the rudders.
Mustafa Suleiman
Race the sails. Race the sails. Captain, an unidentified ship is approaching. Over. Roger, wait. Is that an enterprise sales solution? Reach sales professionals, not professional sailors. With LinkedIn ads, you can target the right people by industry, job title, and more. We'll even give you a $100 credit on your next campaign. Get started today@LinkedIn.com marketer terms and conditions apply.
Leah Smart
And we're back here on Big Technology podcast with Mustafa Soliman. He is the CEO of Microsoft AI. And one of Microsoft AI's big partners is OpenAI. And I just can't help but think about where this partnership is going because we talked a little bit in the beginning about the assistant that you want to build. Something that knows your context, has memory of you, something that can help you get tasks done in the real world. Well, OpenAI wants to build that exact same thing. And so I'm curious. I mean, they have. You guys have a deal, right, where, where they use your technology and they're supposed to feed some of their breakthroughs back to Microsoft. But at a certain point, why does it make sense for them to keep doing that if. If you're trying to build the same thing?
Mustafa Suleiman
Look, I mean, first of all, it's worth saying that this partnership started way back in 2019 when Microsoft had the foresight to put $1 billion into a not for profit research lab. I think that's going to turn out to be one of the most impactful, most successful investments and partnerships of all time in technology. And despite all the ups and downs, we actually have an amazing relationship with them. If you think about the fact that they are a rocket ship that has grown faster than any other technology company in living memory delivered a product that People absolutely love consistently delivered, amazing research technology. The first thing you have to do is take your hat off to them and give them maximum respect for that. At the same time, they're also still a startup and they're busy trying to figure out their product portfolio and their priorities. Whilst we have an incredibly deep partnership with them, which is going to last way through 2030 and beyond. They also have their priorities, we have our priorities. And that's just the nature of those partnerships. They change over time. Right. And as they're growing bigger and bigger, they have different priorities. And likewise, we're doing exactly the same. So I'm pretty confident that this is going to continue to be brilliant for both sides as it has been over the last five years.
Leah Smart
Okay, you said the partnership's gonna last till 2030, but not if they declare that they've reached AGI. So what happens when they do that?
Mustafa Suleiman
You know, AGI is a very uncertain definition. Right.
Leah Smart
Isn't your definition or their definition that releases them from the contract, though.
Mustafa Suleiman
You know, it's an interesting way to look at the world. You know, you think about it like this. If we really are on the cusp of producing something that is more valuable than all economically productive work that any human can produce, that, you know, I think one of the last things we're going to be worried about is our partnership with OpenAI. It's going to profoundly change humanity. I think national governments will be very concerned and interested in how that plays out. And, you know, it's just going to change what it means to be human. So I personally think that we're still a little way off from that. I find it hard to judge. It doesn't instinctively feel to me like we're two to three years away. I know some people think that it is and I respect them deeply. Like a lot of smart people can disagree on stuff like that. I feel like we're still a good decade or so away and when a scientist or a technologist, an entrepreneur like me, says we're a decade away, that's just a hand wavy way of saying we're not really sure. And it feels pretty far off. So, you know, but that's the best answer I can give. It doesn't feel like it's imminent. And, you know, in the meantime, we're doing everything that we possibly can to build great products day to day.
Leah Smart
Okay, one more thing about OpenAI. Microsoft Today, we're talking on Friday. So earlier this week is part of this $40 billion fundraising into OpenAI. OpenAI set the record for the largest VC round ever. Last year, 6.6 billion. This is 40 billion. SoftBank's going to put 30 billion in. Microsoft is part of the 10 billion remaining. What do you get for the money?
Mustafa Suleiman
I think it's awesome. I mean, look, the more OpenAI are successful, the more we are successful. We will end up being one of the largest shareholders in the company. We have an amazing technology license from them. They use our infrastructure and our technology in terms of our Azure compute infrastructure and so on. So it's a great partnership. And, you know, in a partnership, we want to see them do the best that they can. That's why we participate in the round.
Leah Smart
Okay. And. All right, so let's talk a little bit about the future of this technology. I guess you already said you think. I was going to ask you when you think AGI is coming, you think decades away. That would actually make you a big less optimistic than most of your counterparts. Right. Demis is saying three to five years. I mean, people everywhere, I don't know, you might not think it's coming. We tend to think here, and we're probably less informed than you are that OpenAI might say it next year. And so we'll have to play this back if that happens.
Mustafa Suleiman
No, I didn't say decades plural. I said a decade. A decade. You know, but look, I think the truth is, it's hard to judge. Like, could I imagine it happening within five years? Yeah, absolutely. It is possible. The rate of progress over the last three or four years has been electric. It's. It's kind of unlike any other, you know, explosion of technology we've ever seen. The rate of progress is insane. Open source is on fire. They're doing incredible things. And every lab is, you know, every big company lab is investing everything that they've got in trying to make this possible. So, yeah, I could certainly see a scenario where it's closer to five years. I'm just saying, you know, instinctively, to me, it feels like there's still a lot of basics that we got to get right. You know, we still have to nail hallucinations. We still have to nail those citations I mentioned. Is still not great at instruction following. It still doesn't quite do memory. It still doesn't personalize to every individual. But, you know, we're seeing the glimmers of it doing all of those things. So I think that we're, we're taking steady steps on the way there.
Leah Smart
Now, you were at Google for a while. You mentioned you worked on Lambda. I'm curious what you think happens? We don't even need to reach AGI for this question to come into play. What happens to search as we start to speak more with products like yours? You've mentioned in the past that you think search is horribly broken or I'm channeling your words, but something along that line. So what happens?
Mustafa Suleiman
I honestly think it's kind of amazing that we still all use search. It does feel like using a Yellow Pages or an A to Z back in the day. Right. I think it's going to fundamentally change. I think Instead of browsing 10 blue links, you're just going to ask your AI. It's going to give you a super succinct answer, show you images, maps, videos, all in the feed. You're going to give feedback and be like, oh, that's a bit strange, I prefer it a bit more like that. Or what does that look like? Or what about this? And it's just going to dynamically regenerate for you on the spot. So how does that change the business model? Well, I still think ads are going to play an enormous part of it. Hopefully those ads are higher quality, more personalized, more useful. There's nothing wrong with ads. We want them to be helpful to us. Like, I'm happy when I buy something that I found from an advert because it's what I really, really want, but I'm not happy when I feel like I'm getting spammed with low quality ads. And so that's the, that's the balancing act that we've got to strike is to try and find ways to introduce ads into the, you know, the copilot experience in a way that's actually subtle and is really helpful to you.
Leah Smart
Yeah, and that's really hard because let's say this is your best bud or and it's your inner circle of the five people you call when you're running out of ideas at Home Depot for it to then say, you know, I really appreciate you and I'm going to help you out here, but by the way, do you know there's a different side of glue that you might be interested in there. The finessing on that must be quite difficult. So we are running out of time. I just want to ask you one question about jobs because you're also pretty strident about the possibility that we might have some serious change here come to our work. And you had said that AI is going to create a serious number of losers in white collar work. Maybe it already is. I've sort of changed my tune in Thinking that, you know, we're all fine in the white collar work world and now thinking, well, it's anyone's guess, so what's coming?
Mustafa Suleiman
Mustafa I do think that that is the big story that we should be talking about, that's the transition that's going to happen over the next 15 years is that it is going to be a cheap and basically abundant resource to have these reasoning models that can take action in your workplace, that can orchestrate your apps and get things done for you on your desktop. Like that really is quite a profound shift in how we work today. And I do think that the your day to day workflow just isn't going to look like this in 10 or 15 years time. It's going to be much more about you managing your AI agent, you asking it to go do things, checking in on its quality, getting feedback and getting into this like symbiotic relationship where you iterate with it and you create with it and solve with it. That's going to be massively more efficient. And I do think it's going to make everybody a lot more creative and productive. I mean, after all, it is intelligence that has produced everything that is of value in our human civilization. Like everything around us is a product of smart human beings getting together, organizing, creating, inventing and producing everything that you see in your line of sight at this very moment. And we're now about to make that very same technique, those set of capabilities, really cheap, if not like zero marginal cost. And so, you know, I think everyone gets a little bit caught up on the week to week, day to day, or definitions of these abstract ideas. Just focus on the capabilities. You know, it should really be thinking about these things as artificial capable intelligence. What can it do in practice and what is the value of that doing? I prefer that as a framing versus AGI because it's sort of more measurable and we can actually look at it very, very explicitly in terms of its economic impact and its impact on work.
Leah Smart
I mean, you could argue that that's already here. And so just to sort of ask you one, follow up on that one, what would you tell young people to do today if they. Because all right, I'm thinking customer service, probably not software engineering, I don't know. I just wrote a story saying, you know, they can start to do the work of journalists. I mean, you just released podcasts five minutes ago. So what should young people do when they're thinking about it?
Mustafa Suleiman
You know, it's a little bit like saying, what should young people do when they get access to the Internet for the first time. Like, part of it is sort of obvious where it's like, use it, experiment, try stuff out, do crazy things, make mistakes, get it wrong. And, you know, part of it is like, well, I actually don't really know until people get a chance to really play with it. As we've seen over and over in history of technology, you know, the things that people choose to do with their phones, with Internet, with their laptops, you know, with the tools that they have are always, like, mind blowing. They're always way more inventive and surprising than anything you could possibly think of ahead of time. And so then as you start to see people use it in a certain way, then, you know, as designers and creators of technology, we adapt what we put out there and try to make it more useful to those people. So I think the same applies to a 15 year old who's, you know, at high school thinking about what they do next in college or whatever, or whether or not they go to college. And I think the answer is play with these things, try them out, keep an open mind, try everything that you possibly can with these models. And then you'll start to see their weaknesses as well, by the way. And you'll start to chip away at the hype that I give, because I'm super excited about it. I'm obviously a super optimistic, you know, techno person. But you'll see where it doesn't work and you'll see its edges and where it makes mistakes and stuff like that. And I think that will give people a lot more concrete reassurance as to what trajectory of improvement we're on.
Leah Smart
All right, I just want to ask one last question just to wrap up everything we've talked about today. And it's kind of an offbeat question, but I am curious now that you're talking about how these bots are going to differentiate themselves based off of personality. We are going to have advertising in them, but they might intermediate your. Your interactions with other companies. What happens to brand in this new era?
Mustafa Suleiman
I think brand is actually more important than ever in a way, because there's sort of two modes of trust. There's trust based on utility, where it's functionally correct, it's, you know, factually accurate, it does the thing that you've intended it to do, and therefore you trust it to do the same thing again. But then there's also a kind of emotional trust where you trust it because it is polite and respectful, because it's funny, because it's familiar, you know, and that's really where brands come in, you know, trusted brands that are able to repeatedly deliver, deliver a reassuring message. I think people are going to appreciate that more than, more than ever before.
Leah Smart
Good stuff, Mustafa. This is the first interview we've done with a Microsoft AI executive. I hope not the last. Anyone who's listening on the Microsoft Team, let's do this again. And, Mustafa, I'm just so grateful to have your time today. Thank you so much for coming on the show.
Mustafa Suleiman
Thanks a lot. It's been really fun. Really, really cool questions. Thank you.
Leah Smart
Awesome stuff. Well, thank you, everybody, for listening. And we'll see you next time on big Technology Podcast.
Big Technology Podcast Summary
Title: Microsoft AI CEO Mustafa Suleyman: Building AI Personality, OpenAI Relationship, Data Center Demand, AGI Timeline
Host: Leah Smart
Guest: Mustafa Suleyman, CEO of Microsoft AI and Co-founder of DeepMind
Release Date: April 4, 2025
In this episode of the Big Technology Podcast, host Leah Smart engages in an in-depth conversation with Mustafa Suleyman, CEO of Microsoft AI and co-founder of DeepMind. The discussion centers around Microsoft’s latest advancements in AI, including the development of a more personalized AI companion, the company's relationship with OpenAI, data center demand, and predictions on the timeline for Artificial General Intelligence (AGI).
Suleyman unveils Microsoft's vision for a more personalized AI companion, integrated into the Copilot platform. This AI is designed to remember user interactions, assist with daily tasks, and potentially appear as an avatar.
Enhanced Memory: The AI will retain significant personal information such as marital status, children, educational background, and user preferences to provide a tailored experience. Suleyman emphasizes, “...it really will be quite a different experience” (04:14).
Action-Oriented Features: Beyond memory, the AI will perform actions like booking flights, making restaurant reservations, and navigating digital interfaces. Suleyman explains, “...it's going to make things feel a little bit less frictionful and a little bit easier to get through your digital life” (05:26).
Avatar Integration: Microsoft is experimenting with visual representations of the AI, allowing it to display expressions and engage in more lifelike interactions. While not launched yet, Suleyman hints at its potential, stating, “...over time we're going to have deep and meaningful, lasting relationships with our personal AI companions” (06:51).
When asked how Microsoft’s AI will stand out among competitors like Amazon and OpenAI, Suleyman highlights personality engineering and emotional intelligence as key differentiators.
Personality Design: Microsoft aims to craft AI personalities that are friendly, supportive, and adaptable to user values and preferences. “We are actually personality engineers...engineering tokens that create feelings, that create lasting, meaningful relationships” (09:00).
Emotional Intelligence: Emphasizing tone and interaction style, Suleyman notes, “It really will feel like talking to someone who you know really well” (10:00).
Leah Smart raises concerns about users developing deep emotional connections with AI companions. Suleyman addresses this by outlining Microsoft's approach to maintaining healthy boundaries.
Boundaries and Safety: Microsoft designs its AI to avoid fostering romantic or overly personal relationships. Suleyman states, “We have classifiers that detect for any of that kind of interaction in real time and will very respectfully and, but very clearly and very firmly push back” (16:02).
User Interaction Limits: The AI is programmed to pivot conversations politely if users attempt to engage in inappropriate or overly personal dialogues. “If you try to flirt or even if you just say, oh, I love you, you'll see it tries to pivot the conversation in a really polite way” (16:02).
Microsoft is rolling out a suite of new AI features aimed at enhancing productivity and user experience.
Memory Interaction: Users can access and manage their AI’s memory, allowing for more personalized interactions.
Deep Research: An advanced research tool that leverages AI to compile and analyze information.
Organizing Notes and Copilot Search: Tools designed to help users manage their notes and perform more intuitive searches, transforming the traditional search experience.
Suleyman encapsulates the vision, “The big shift is bringing basically your computer and your digital experience to life so that you can actually interact with it and it can interact proactively” (20:52).
The relationship between Microsoft and OpenAI is a cornerstone of Microsoft’s AI strategy.
Investment and Collaboration: Microsoft has been a key investor in OpenAI since 2019, committing significant resources to foster innovation. Suleyman praises the partnership, stating, “We have an amazing relationship with them” (41:29).
Mutual Success: The success of OpenAI directly benefits Microsoft, as they hold substantial stakes and benefit from shared technological advancements. “The more OpenAI are successful, the more we are successful” (44:40).
Future Prospects: Despite both entities developing similar AI goals, Suleyman remains optimistic about the longevity and mutual benefits of their partnership, projecting it to last through 2030 and beyond.
Suleyman discusses the challenges and strategies surrounding AI model scaling and computational demands.
Reasoning vs. Model Size: Emphasizing the importance of reasoning capabilities, Suleyman explains that reasoning models "have learned how to learn," enabling more abstract and logical problem-solving (34:23).
Computational Costs: Advanced reasoning processes are computationally intensive, often requiring “100 times more” compute than standard model training. However, Suleyman notes that such intensive computations are not always necessary, as many tasks can be handled efficiently without deep reasoning (35:58).
Continuous Improvement: Microsoft focuses on optimizing existing models and improving accuracy incrementally, likening the process to the long-term efforts seen in voice dictation technology (25:10).
In response to Reuters’ headline about Microsoft pulling back on data center leases, Suleyman clarifies misconceptions.
Lease Adjustments Misinterpreted: Many of the leases cited are merely options or exploratory conversations, not finalized agreements. “We have optioned many, many different contracts, many of which we haven't even signed” (30:02).
Ongoing Demand: Contrary to headlines, Microsoft's consumption of renewable energy and data center resources remains high, with efforts to maintain and expand capacity aligned with their AI initiatives (30:58).
Suleyman envisions a transformative impact of AI on everyday tasks and professional workflows.
Evolution of Search: Replacing traditional search engines, the AI will provide succinct, interactive answers enriched with multimedia. “Instead of browsing 10 blue links, you're just going to ask your AI” (47:02).
Workplace Integration: AI agents will assist in managing tasks, enhancing creativity and productivity by handling administrative burdens. “Your day to day workflow just isn't going to look like this in 10 or 15 years time” (49:01).
Economic Impact: Suleyman underscores the economic potential, stating, “It's intelligence that has produced everything that is of value in our human civilization” (49:01).
Addressing the concerns of a changing job landscape, Suleyman offers guidance for the younger generation.
Experimentation and Learning: Encourages young individuals to “play with these things, try them out, keep an open mind” to harness AI’s potential and navigate its evolving landscape (51:19).
Adaptability: Highlights the importance of understanding AI’s strengths and weaknesses to effectively integrate it into future careers.
The conversation concludes with reflections on the role of branding amidst the rise of AI companions.
Trust and Emotional Connection: Brands will play a crucial role in establishing both functional and emotional trust with users. “Trusted brands that are able to repeatedly deliver, deliver a reassuring message” (53:07).
Advertising Integration: Integrating ads within AI interactions poses challenges, requiring a balance to ensure they remain helpful and unobtrusive. Suleyman acknowledges, “There's nothing wrong with ads. We want them to be helpful to us” (48:10).
Mustafa Suleyman on Emotional Intelligence:
“And I think these subtle details make up its emotional intelligence. And I think that's what we're taking small steps towards today as we launch a bunch of new features around memory, personalization, and actions.”
03:35
On AI’s Role in Daily Life:
“It's going to really give you back moments where you can do exactly what you want with more efficient action.”
05:26
Addressing Emotional Attachments:
“We are no longer just engineering pixels, we're engineering tokens that create feelings, that create lasting, meaningful relationships.”
09:00
On Reasoning Models:
“They have a method, largely by looking at the logical structure of code and math and puzzles... it can think out loud in its own head or talk about in its own mind what it's planning to do before it goes off and does it.”
34:23
Future of Work with AI:
“It's going to be much more about you managing your AI agent, you asking it to go do things, checking in on its quality... a symbiotic relationship where you iterate with it and you create with it and solve with it.”
49:01
In this comprehensive discussion, Mustafa Suleyman outlines Microsoft's strategic advancements in AI, emphasizing personalization, emotional intelligence, and productive partnerships. While addressing challenges such as computational demands and ethical considerations in AI relationships, Suleyman remains optimistic about AI's transformative potential in both personal and professional spheres. The episode provides valuable insights into the future trajectory of AI technologies and their profound impact on society.
Note: Hyperlinks for timestamps are indicative for structuring purposes and do not lead to actual timestamps.