Transcript
A (0:01)
This is the Everyday AI show, the everyday podcast where we simplify AI and bring its power to your fingertips. Listen daily for practical advice to boost your career, business and everyday life.
B (0:16)
You shouldn't use as many AI tools as I do. It's actually a recipe for disaster. And I think that shiny object AI syndrome is one of the biggest problems in today's enterprise landscape. But although I don't think most business leaders should be using 20/AI tools or systems every week, I do think it's imperative to understand the landscape. And obviously it's ever changing because in just about any field, sector or modality, there's probably a juggernaut unicorn AI company that has created a top notch AI tool that's redefining work in that given space. So on today's show, I'm going to give you a lay of the AI land because I've been there, I've wasted thousands of hours so you can just learn from me as I tell you what types of tools are good for what reason. So on today's show and Volume 16 of the Start Here series, we're going to walk from chatbots to AI agents and everything in between, and quickly and simply lay out the 11 different AI parent categories where tens of thousands of AI tools mostly all fall under. Because, yes, there's a lot more to AI than chatbots and agents. So let's just go ahead and connect all the different dots and categories in between. All right, let's get into it. If you're new here. This is Everyday AI and our Start Here series. So here's the big picture of what you need to know for today. Well, there's hundreds of AI tools literally launching every single week. And most people feel overwhelmed. Well, I do too. I try to keep up with it. And I don't recommend you try to keep up with the hundreds or thousands of tools in the AI space that are released every week because most of them are, well, kind of garbage and don't hold a lot of utility. But almost every single one of those tools falls under, you know, maybe 11 of these parent categories that we've created for you today. And there's actually very few exceptions. And yes, some of these categories are extremely broad. But I did that for a reason. And I think once you understand these 11 categories, you can kind of stop chasing tools and start building a stack that makes sense for you. Because like I said, even within all of these individual categories, the functionality is changing constantly. So, you know, you might say, oh, there's 50 categories, but they're Changing so much and adding so many new features and functions. Well, it's, I think it's easier to just understand just a couple less than a dozen of these categories. So this episode is your map to every type of AI tool that exists right now. So on today's show, stick with me for the next 20ish minutes and you're going to learn why. Well, well, why? You're probably only using one or two of these AI categories and you're probably falling behind because of it. And again, you shouldn't use tools from all of them. You're also going to learn the single framework that makes sense of every AI tool from chat GPT to cursor to Suno, and how to build a personal AI stack that matches your job. Your competitors. Find out. All right, welcome to our Start Here series. This is the Everyday AI Essential podcast series to both learn the AI basics and to double down on your AI knowledge. I created this series well because I didn't have a good answer when so many people said there's like 700 plus episodes. Where do I start? Well, you start here with the Start Here series. If you are picking this up midway through, that's okay. But I highly recommend you listen to all of these episodes in order. They're shortish, they average like 29 minutes or something like that. And you can go to start here series.com that's going to give you free access to our inner Circle community. And you can go check out our Start Here series channel in there where you can go listen, read and just connect with other people who are going through this series in order. So if you have any questions, make sure to hit me up there after you join the inner circle community. And if you missed our last Start Here series episode, we talked about how everything is fake and how your company can leverage human expertise properly and fight AI work slot. That one was a good one. It's actually fun one to do. All right, but let's get into the road map here in the 11 categories. Explain. Here they are right away. Number one, text reasoning assistance. Number two, multimodal AI platforms. Number three, AI search and research. Number four, voice and speech AI. Number five, image generation. Number six, video generation. Number seven, music generation. Number eight, design and visual content. Number nine, vibe coding app builders. Number 10, AI coding tools. And number 11, AI agents. All right, and let me just tell you this. I took a little bit of liberty and played around with these how to categorize all these different tools, right? Because as an example, something like Gemini, well, they fit in Basically all of these. Right? And in theory, we could just have combined number one, two and three into large language models. But there are literally thousands of actually pretty good tools right out there. Right. Even though I think that 99% of them are pretty garbage. There's literally hundreds of thousands of AI tools. But so there's actually thousands of great ones that fit in to these 11 different categories. And so I did kind of take a little bit of leeway and maybe this will make sense as we unpack this a little bit more. But you might be wondering, well, what's the difference between a text reasoning assistant and a multimodal AI platform? So, yes, in the beginning you might just be saying, well, Isn't that just chatgpt or can't Gemini do you know 9 out of these 11 categories fairly well? Well, yes. But as you'll see as we unpack this a little bit more, there are some unique tools that only fit in maybe one of these categories. And I think as we start to think about how you can practically apply these different categories and kind of create that, you know, essential kind of tool category stack, it's important to understand the differences. Yes, there's crossover, but I think that we need to kind of properly go through this. So don't forget, AI is not new, all right? And I know if you're a longtime listener, you're like, okay, Jordan, I get it, right? But I'm trying to keep this very basic and I think this is going to be one of those episodes that a lot of people listen to and I'm going to point people toward because, yes, there's. There's always an AI tool that can do something, right? And Even in these 11 different categories, all of these also have sector specific tools that really shine, right? So as an example, there's, you know, even image generation, there's image generation tools that are essentially just wrappers of, you know, Google's Nano banana or chat GBT's image tool for different sectors, right? So also keep that in mind as you're using these different tools. A lot of them are just based off of, well, one of the core platforms from one of the big players. Not all of them, but a lot of them are. So also keep that in mind. But also keep in mind AI is not new, right? I want to give a very, very quick kind of recap of the last 60 years. This will be the fastest ever, right? But AI has been around since the 50s and the first AI chap out was actually Eliza in the 60s, and kind of the big boom of AI was actually in the 80s, right? A lot of people think it started with Chat gbt. No, you had expert systems in the eight, but essentially these were very rigid if then else logic pieces of, you know, old school AI algorithms. And they weren't really meant for general use cases, right. A lot of them as an example, you know, you could look at an old school artificial intelligence in banking and all it was, it was a very complicated rules based decision tree, right. That when maybe an inquiry came into one of those old computer systems that were, you know, the size of a school bus and literally only did one thing very slowly, all they did was kind of traverse this path of this very rigid if then else logic. And usually if there's one thing wrong, right, one extra comma, one extra space, the entire thing broke. And that's obviously very different from today's generative AI, right? And a lot of that happened because of a very famous paper essentially from Google called attention is all you need. And they're 2017 Transformer architecture. And that really unlocks well AI that's not as rigid and that's what you know, kind of Chat GPT exploded this scene which is generative AI. So you know, when we had for decades AI that was very rigid. Now it's moved into gen generative reasoning and intuition. And I think that ChatGPT really helped launch this new category. Yes, there were other generative AI tools before ChatGPT, you know, maybe by a year or two. But I do think that it was the surgeons of ChatGPT that popularized AI tools, right? It brought in a lot of funding that really helped pave a lot of these different categories in all the different tool kind of parent umbrellas that we're going to be going over today. So let's talk about category one. That's just text reasoning assistance. This is you type a category, you type a query into chat GBT, into you know, anthropic Claude, into Gemini, you type in text, you get text out and well, who is this for? It's for everyone. This is kind of the gateway drug of AI, right? But it's important to know by default everything is reasoning now. So if you maybe were a Heavier user of AI in 2022, 2023 and you're like this thing stinks, I'm never using it again. And you're just waking up now today. Yeah, the models are a lot better. They are smarter than most human experts when it comes to creating economically viable work. But they do reason by default, they think by default, right? So Some of the models are slower and I also. Right, we've gone over this multiple times in the Start Here series, especially when you are using tools in category one, which I think is where a lot of people spend time. Do not sacrifice speed for using the right freaking tool. Right? Don't use like an instant version or a non thinking version of any of these tools. Right? We literally have. AI moves too fast to follow, but you're expected to keep up. Otherwise your career or company might lag behind while AI native competitors leap ahead. But you don't have 10 hours a day to understand it all. That's what I do for you. But after 700 plus episodes of everyday AI, the most common questions I get is where do I start? That's why we created the Start Here series, an ongoing podcast series of more than a dozen episodes you can listen to in order. It covers the AI basics for beginners and sharpens the skills of AI champions pushing their companies forward. In the ongoing series, we explain complex trends in simple language that you can turn into action. There's three ways to jump in. Number one, go scroll back to the first one in episode 691. Number two, tap the link in your show notes at any time for the Start Here series. Or you can just go to starthereseries.com which also gives you free access to our inner circle community where you can connect with other business leaders doing the same. The Start Here series will slow down the pace of AI so you can get ahead. Insanely helpful technology that can create outputs, deliverables, artifacts that are indistinguishable from human experts. Yet so many people aren't getting that just because they're impatient, right? And they don't want to maybe wait 3, 4, 5 minutes for a response and instead will, well, they'll just say, I'm just going to get this fast one, or I'm going to use this model that doesn't think or doesn't take as long. Number one, stop doing that. Right? There's a reason why the people who are getting the most out of these models are, you know, usually have multiple screens and, you know, multiple platforms. Because you have to wait. You know, you have to wait. You have to be patient, you have to have a good workflow. This is why I usually recommend, yes, pick your AI operating system of choice, but have backups, right? Or be working on multiple projects at once. You do have to be a little bit, you know, okay at multitasking and bouncing between different projects because you should be taking your time when working in category one, which is our text reasoning assistance. All right, Category two, this is multimodal AI platforms. And for the most part it's the same quote unquote tools in category one as they are in category two. And I think it's actually important to differentiate these use cases because I would say the overwhelming majority of the billions of people that use generative AI every single week are using these same tools, but they're only using them for, for the most part, category one text reasoning, they're not using them as a multimodal AI platform. All right, so what that means is, well, multimodal in, multimodal out. All right, so the main players here are the same as the main players in number one. So that's ChatGPT, Gemini, Claude, also Microsoft Copilot, and they really are expanding into multimodal operating systems. So who is this category for? Well, anyone that wants one AI platform and across multiple modalities. So I think Google is by far the leader in this space. And as we go on, you'll probably see Google Tools mentioned more than any of the others because they are absolutely dominating in multimodal AI. But what does it mean for a multimodal AI AI platform, let's say Google. Right. A lot of people don't know. In Google's AI studio, which is a kind of more customizable or, you know, developer friendly version of Google Gemini, you can upload videos, right? And Google Gemini 3.1 Pro doesn't just, you know, transcribe videos, it can see them, right? Yes, it's very token intensive. You know, you might burn through a little bit of your budget, but it can actually see and understand videos. So that's the difference between and why I think it's important to have two different categories for a literal multimodal input. And a lot of people still, you know, don't take advantage of that in a chat. GPT, Claude copilot, right, that it can, you know, upload screenshots, you can, you know, output, you know, depending on which system. Right. You can output code, you can output graphics, you can output images. In some, you can output video. So it is important, even though a lot of people spend most of their time in category one, those the three or four big boys in the room, they play in category one and two. Which also leads me to. Yes, I know, I said we could in theory combine categories one through three, because category three is AI Search and Deep research. And you have a lot of the same players, but they are kind of different products. Right? They have their Kind of dedicated, you know, features or modes. So in ChatGPT you have deep Research. In Gemini you have Deep Research. In Claude you have extended, kind of the extended reasoning kind of version of this to take longer and do deeper research. And there is also a research tab inside Claude. So it is kind of a different platform, but then you have a completely new set of players in here as well. One of the ones that most people are probably most familiar with is, is Perplexity. You also have, you know, Notebook lm, I think technically fits in this category as well. Although, you know, there's completely so many different categories where you could place notebook lm you have tools like, you know, u.com and then you have literally hundreds of great industry or sector specific researchers as well. So who uses this? Well, I think your average everyday professional that just needs to verify facts, but also researchers, analysts, strategists, and like I said, some of the more, you know, especially on the academic research side, legal research side, some of those tools are probably going to be a little bit more familiar for people working in those sectors. So I'm not going to confuse everyone with all those names. Right? The Harvey of the legal worlds. You know, there's so many on the academic side, but essentially AI you need to understand and I think people always skip over this. The source of your answer is extremely important. So when you talk about large language models having AI, search and deep research, it's, it's, it's actually not quite a novel concept. They've been doing this for multiple years. But you have to realize there's really three different main sources for where a large language model gives you its answers. Right. Number one is its own internal training data. So as an example, if you're using an offline, you know, model that you downloaded and you're not connected to the Internet, which a lot of people now are doing as we're moving into, you know, more powerful computers, more capable models, things like openclaw, where people are, you know, downloading open source models and running them locally. Those models are running off of usually very old training data. So you have old training data, you have data that you can upload, right. Both individually in chats, via connectors and integrations, you know, via projects, GPTs, etc. So there's training data, data that you can upload or connect via, you know, different mechanisms within these large language models. And then there's the ability to browse the web. So you really have to have a handle and an understanding and make sure you're always checking the summarized chain of thought to see and ensure that they're looking at the right types of websites and they're not making things up or hallucinating anything. And I think this matters because I think humans aren't really going to be using the Internet a lot in the future. That might sound weird. I find myself personally using the Internet less and less every day and I'm spending more and more time in these different platforms. Right, like your perplexity notebook, LM, you know, Gemini's deep research, ChatGPT, deep research etc and I do think that as agents become more prevalent, it's just going to be agents really browsing the web. That's why we have now all these agentic protocols that are making websites more accessible and well, the data in the websites more readily available to agents. Right. I could see a day where, you know, obviously we're already starting to see it, you know, even from a technical kind of SEO or AIEO or geo, whatever you want to call the new SEO for AI, right, where there's multiple versions of a website, maybe there's one version that's for humans and one version that's for AIs. And probably the version for AIs is ultimately going to become way more important for most businesses that rely on discoverability online. All right, category, category four, another big one here and it's, it's an emerging one and I think a trending one in 2026. And this is voice and speech AI. So this is both speech to text. It is text to speech. So yeah, this is, if you didn't know, yeah, there's an AI for that and they're good. Right. I've even been building a secret little project in this category and even the open source models, right, so not just the frontier models that you're paying for, but this I think is going to be really important moving forward. Not saying that people aren't going to be able to type or that everyone wants a text to speech, voice narrating something. But you know, I usually, I usually see where I'm spending more and more of my time and I'm kind of forecasting that out and saying, okay, I think a lot more people are going to be using this because like I said, I use hundreds of AI tools every year and I've used thousands of them in the three plus years I've been doing everyday AI. So you know, I, I fancy myself the average knowledge worker, right. I'm not overly technical, I'm fairly, you know, technical. So I'm kind of in the Middle, Right. So I think things that, at least for me personally, stick, it makes sense. I think that they will make sense for many other people. And I think that's why voice has really been a trending category. So some of the main players here, 11 Labs Mistral, actually has a great new model that they announced, which is Voxtral Fireflies, you know, for, you know, AI meetings. You know, you have a lot of those within Copilot, within Google Gemini. They have their own meeting assistant that transcribes meetings. You have granola, you have otter, right? There's so many big players, both on the text to speech and on the speech to text side. So who uses these? Well, anything from podcasters like myself, right? Video creators, global teams, corporate training, accessibility. There's so many different use cases. But one of the reasons why I think this category is going to become increasingly more and more important is there is a literal knowledge layoff coming, right? So both as companies are, unfortunately, we're seeing because of AI laying off hundreds or thousands of workers for some of the bigger tech companies, I do think that's going to unfortunately sprinkle into kind of corporate America throughout the second half of 2026 and into 2027. So I think you are not only companies are losing a lot of their institutional knowledge when they're laying people off, right? They're. They're doing it public companies are doing it because it helps their bottom line, right? And they can become leaner. And a lot of people don't know what the future AI jobs are going to look like. So they're like, well, we might as well get rid of the old jobs now, save up money. And so when we need those new AI jobs, we can go hire people in. Well, we'll be leaner in the process. But you also have the Silver Tsunami, which I talk about a lot, right? You have more and more people who are retiring, and a lot of these more senior workers maybe have decades of institutional knowledge, subject matter expertise in their head. And I think that's another reason why some of these platforms are becoming more and more important, right? Documenting this institutional knowledge, right. Doing zoom meetings and just talking to senior people, right? That maybe they're leaving in a couple of years, you know, trying to understand their thought process and then using that as well. Your company's internal ip. I think these type of tools are extremely important because. And now it's also very common for, you know, it's starting more in the coding space, but it's more and more common for people to just be talking Instead of typing, right. As large language models become better at understanding natural language, even for me, I find myself, I'm a decent typer, right? I'm in front of the computer 1012 hours every single day. Large language models are very good now, I think at understanding natural voice. And you can speak about four times as fast as you can type for the most part, right? So this is another reason why this category. And then on the flip side, right, text, speech. I know this one is maybe me more personal. I hate reading. I hate reading huge lines of verbose AI generated text, right? Because that's where I spend the majority of my day, right. So everything I read for the most part is AI generated and extremely long, even with custom instructions and, you know, certain settings, right. So I think, you know, being able to use these text to speech tools or, you know, they're baked into a lot of, you know, ChatGPT and Gemini as an example, you can just click the read aloud button. It's, it's a big category in and of itself. All right, number five, AI image generation. So how this works, it's a little different. So for the most part, most of the things that we've talked about so far in the first categories were all based off some of that original, you know, transformer models, right. The original research from Google, you know, you can say it's next token prediction, but like on steroids, AI image generation is a little bit different. So these are something called diffusion models. So essentially they start with random noise and they've been trained on a large data set and then it kind of refines it into coherent images. So you could kind of think of it like, you know, I don't know if you have a little petri dish of milk and you, you know, drop some black ink in it and it eventually diffuses into a photo, right? That's kind of what it does behind the scenes. I actually like, I really enjoyed the older versions of like, you know, OpenAI's Dolly when it was like really slow because you could actually kind of see, right. In some of the earlier Mid Journey versions, right. When the images are real slow, you could understand the diffusion process because you could see it and watch it, right. One image might take a couple minutes and it turned out absolutely terrible, right. Like back in the dolly two days, but then you could actually understand and see. Okay, I kind of understand what this model is trying to do. All right, so some of the main players here, Mid Journey was one of the OGs, even though I don't think that they're any longer like a top three name, but you have to kind of tip the cap to them. Then you have Flux Ideogram and then as always, you have the big players, right? Which is GPT Image 1.5 from OpenAI, Nanobanana Pro, Nanobanana Pro 2 and then also Microsoft has recently announced some pretty decent, right? Some top five esque AI image models. So who uses these? Well, I think anyone can, right? I think especially with using something like Nano Banana Pro that can make slides. All right, we're going to kind of get into that category, a dedicated category, right? But you know, things like GPT Image 1.5 inside of ChatGPT, Nano Banana Pro inside of Google Gemini, being able to make infographics, things like that. So yes, I think people think that these are only for designers, marketers, content creators, but I think they can really be for anyone, right? Because we also have to understand, right? I say this as someone, I'll say mid career, right? I think the younger generation doesn't really care about text, right? They care about videos. And that's also the base for creating AI videos is both AI image generation, but also just more interactive graphics and just better visual elements. I think businesses should be starting to experiment not just on social media, on their website, on traditional marketing materials as well that are just, you know, blobs of thousands of words of text. This is a great, a great kind of tool category to start exploring if you haven't already. Because I think standalone image tools are losing ground, right? As now we have these AI image generators that are really sweeping the field. All right, Category six, this is AI video generation. So here's how it works. Well, it's kind of like the in image generation but. But then like times 60 for all the different, you know, frames per second or you know, 30 frames per second. So it's slightly similar technology as they have these diffusion models, but they are also then extended into the time dimension and then there's denoising frame sequences. The other thing with good video models is they are extremely expensive to run, right? So as an example, Sora, right, Depending on when you're listening to this. But so technically it was just last week opening AI shut down. Sora. One of the reasons why. Well, video generators are extremely expensive because they also understand the world, right? I do think that video generation in world model generation are going to start to blend kind of like what we've seen out of Runway, because the video generate generators, good ones, they also have to understand gravity, they have to understand, you know, shadows, reflections, you know, sun, right. Like they have to understand things, you know, character consistency. Right. Like if, if it's a video generation of, you know, someone walking toward a pencil on the ground, but from the pencil's point of view, right. The object is going to shame. Change, shape, dimension, all those things in real time. So the video generation and I think the, the gap that was closed in 2025 because at the beginning of 2025, AI video was not good at all. Right. And then fast forward to the end of 2025 and AI video is actually to the average eye at the end of 2025, indistinguishable for the most part, if it's done well. And I think where we're at now in, you know, mid-2026 is even better right now. It's even people like me, right. I have a background in videography. It's getting harder for me to even realize. So some of the main players Here, you have 031, you have Runway gen 45, you have cling 3O, you have seed Dance, right? See dance gone. It's gone mega viral recently because they're kind of making Hollywood videos and kind of got in trouble. You have Pika 2.5. So who uses these? Obviously, if you are working in or around video social media, I think these are great tools to use and I think that video is one of the hottest and biggest growing categories of all of them. All right, moving into the next category, a category that's actually not super competitive and I'd say it's kind of one of the last defensible creative AI frontiers right now. And that is category seven music generation. So this works a little bit differently than some of the other models. And I won't even get into it the technical side because it technically uses like image generation to visualize what music looks like, but it is for the front end user how it works. Not to too worried about technically how it works under the hood for this one. But you can essentially in text describe the genre, the mood, the lyrics, and then you can receive a fully mixed song in seconds. So some of the main players here. Well, you have Suno V55, which just came out. Yes, I know. Depending on when you're listening to, I don't know, maybe we're on Suno V7 by now. But right now, at the time of this recording, Suno V55, you have UDO, you have 11 Labs now. So 11 Labs was originally a big text to speech player. Now they're in the AI music space as well. Google yeah, like I said, Google in like almost every category because they're a leader in almost every single category. Their new Lyria 3 Pro actually really good being able to create three minute songs and you don't even need a separate subscription, so there's a lot. So you know who uses this again? Content creators. Anyone that's looking for, you know, music to accompany a video, if anyone's needing royalty free audio. And I think that again, there's actually a shockingly a lot of money that these companies are making. I think the last I saw that Suno was at something like $300 million in revenue. So yeah, these companies, even if you haven't heard of them, they are really big, they are really good. And the, the cool thing, right? Because I've always been kind of impressed. I've used Suno since the very early days and they're the biggest player in the space by far. We had their CEO on the show, you know, many years ago. I say many years ago, it feels like like ten, but it was like two and a half. They've been the platform that's been on fire. Right? But you can literally describe anything. You can pull out, you know, just the guitar and you can say, hey, just change the guitar on this. But keep everything else the same. Now you can experiment with your own voice, right? So there's so many cool things that you can do. So you know, maybe as you're understanding these categories a little bit more, maybe there's some more use cases for your company that you maybe didn't think of before. All right, let's get to category eight. And that is design and visual content. This. I know I could have technically broken this one down into like three or four separate categories, but instead I decided to keep it a big parent category because I do think that like, as an example with like Notebook lm and you know, we're going to see a lot of this in Google with Nano Banana, the lines on this are really going to start to blur. Like as an example, if you look at Nano Banana, yes, originally it was okay. It, it uses, it's, it's for photos and infographics, right? But then all of a sudden it's doing complete slide decks and then it's the base for video slideshows, right? But so this category, it is AI that combines content generation, layout, design and imagery into finished artifacts, whatever, whatever that artifact may be. So some of the big players here were, you know, kind of earlier in the game and they're doing things like decks, right? So obviously I Think you now have Microsoft Copilot, which is pretty good in creating decks inside PowerPoint. Actually, I was surprised how good the new task feature in Microsoft Copilot was as at creating decks, right? So even outside of the tools that we talked about, you also have to understand, I'm not mentioning, you know, chatgpt and Quad for each of these because they can all create slides, right? But I think this is a little bit different because Gamma. Gamma is one of the big players in this space. You know, very unique tool that can create, you know, a 20 page slide deck can create, you know, websites, it can create anything visually. And I think it is a really, does a really good job. Similarly, you know, canva's AI Magic Suite falls in that category. Beautiful AI, but then you have a whole nother kind of genre of, you know, AI design and visual content, which would be AI, AI avatars or digital twins, right? So you can upload your, you know, they're actually fairly easy to make, way easier than they were like two or three years ago with tools like Synthesia, Hey Gen, etc, But then you also have other platforms that are, you know, visual content, such as Google Stitch, right? So it helps you and you know, Figma AI, right. Like there's so many, so many tools that fall under here. But, you know, there's just tools that you can type in a sentence and it will create an entire app layout and then you can export that to, well, something in category nine. All right, as we move on to category nine and that is Vibe coding app builders. So here's how it works again, natural language, are you seeing the trend here, y'? All? For the most part, natural language is a very important skill set, right? Because for the most part, this is a starting point for all of our categories is just natural language. So with Vibe coding, right, if you slept through 2025 and missed vibe coding, I think being the word of the year for like Dictionary.com or something like that, you know, you describe an app or in plain English and it can get deployed with a live URL. So again, between 9 and 10 categories, 9 and 10, the, you know, vibe coding app builders and the AI coding co pilots, there's a lot of crossover, similar capabilities, right? Google AI Studio could be in either. But I think the main players that are specifically in the Vibe coding app builders would be like lovable replit bolt V0 by Vercel Base 44. Right? There's, there's a lot of, of ones that are really just made to be a complete infrastructure for an Entire app. Right. I'm going to talk a little bit why I even separated vibe coding app builders into AI coding co pilots. Because yes, they do have a lot of similarities, but they're definitely different. Right. So I'd say people using this are more of the non technical people. So you know, non technical founders, entrepreneurs, marketers who are building tools or anyone with an app idea. Right. They are great platforms. I think. You know, earlier on in the, you know, 2024, early 2025, I wasn't a big fan of these tools, to be honest. Right. But now they're much, they're much better right now that they have complete, you know, front end, back end, auth user management. Whereas before they were just, you know, you could make disposable apps and they were kind of disposable. Right. Very similar to what you can make inside ChatGPT, Claude Gemini. Right. But now they have turned into kind of fully functioning tools. So I think this is one of the most heavily funded categories of 2026 and it's rewriting who builds software and how it's used. Yeah, like literally, if you look at the US stock market, traditional software has gotten squashed from category nine and category 10. So, so great segue to category 10, which is our AI coding co pilots and agents. So here's how it works. Well, AI just can understand your entire code base. It can write code, test it and ship new features. But it is not just coding, it is also any work that can be done on your desktop. Right. Because there's this magical thing that, you know, most people, unless you're a dork like me, I've been playing around with different terminal tools for 20 years. Right. But computers themselves are controlled by something called a terminal. Right. So you can run any command on a computer, including accessing local files on your desktop, creating files, running code, all of those things. So that's why you kind of have a little difference in the kind of vibe coding tools that are ultimately about like, okay, you're creating a piece of software where AI coding is a little different. They can create software and yeah, you can vibe code in all of these tools. I do all the time because the main players here, you have Cursor, Claude Code, GitHub, Copilot, Windserve, Codex from OpenAI, Anti Gravity from Google. Right. But I think that we're starting to see it, these main players also dip their toe into non technical knowledge work as well and just becoming agents in and of themselves. So who uses this? Well, professional software developers, engineering teams at every Level. But like I said, I think with Claude code and Claude cowork, which I think I have on the next one here and I think Codex as well, I think more and more, and I know you've heard me say this a lot, and if you are not a technical person, if you're not a software engineer, if you're not a dev, you don't fancy yourself any of those things. You need to be using Claude code, you need to be using Claude code work, you need to be using Codex. You should probably start using Anti gravity, right? Because as we start talking about the future agentic layer, which right now is, well, computer using agents. But I think that's not the ultimate or the end layer. But I think you really have to understand these tools. Right? And we're going to get to kind of my takeaway advice on that here in a minute. But every single coding tool right now is racing toward fully autonomous agent driven development. Which is why I think if you understand AI tools, right, yes, there's great agent capabilities that can happen in front end AI chatbots, but you don't have every single program, every single file, every single app that you use. You can't connect those. Right. A lot of those maybe live on your local machine. Which is why I think it's important to pay attention to the AI coding co pilots and agents category. It is important. And then last but definitely not least, and yeah, this one could literally be seven different categories. I've actually broken down AI agents into the seven different types of agents. But for ease of today's episode and understanding all the different AI software, not just agentic software, we're going to say category 11 is autonomous AI agents and agentic browsers. So this is how it works. Well, you state a goal and the AI plans the steps. It uses real tools and it completes the work. Right. So there's a lot of main players. All right. And you know, we already mentioned some that you might think were in this like replit, you know, maybe Claude code. But I'm saying for the most part these are more general agents or non technical, non dev type agents. So players like Manus, Genspark, obviously Openclaw definitely fits in this ChatGPT's Atlas browser, perplexity Comets browser, their Perplexity computer, personal computer, Claude Cowork, especially their new computer use that can click, use a mouse, can open different files, different programs on your computer, which makes it really unique. You have Microsoft's new cowork, you know, you have Claude Dispatch. There's so many, right. And I think I would even probably throw Codex into this platform as well or into this category. But this is, I, I'd say for more technically proficient knowledge workers, right, who are looking to delegate as many of their time consuming, mundane tasks as possible. All right, and, and well, I think the main reason why this matters is the shift is happening right now, right? Where AI is no longer just something that you ask a question to and it works reactively. Right now agents are working proactively on schedules, right? Like if you're an old school, you know, technical person, right. Heard about, you know, cron jobs and you know, rpa, robotic process automation. Right? Now that's what AI agents are doing, right? They're running computers on schedules and they can do anything that us humans can do if you're using the right tool at the right time for the right purposes. All right, so Those are the 11 categories. But I want to leave you with this. I want to leave you with some advice. And you know, we talked about, you know, why if you're using only one of the categories, you're probably failing and, you know, a framework to make sense of every tool and how to kind of build the right AI stack. Right. So here's some pieces, pieces of advice as we wrap up. So Gemini is an absolute beast, right. We've actually seen in the recent weeks OpenAI intentionally stepping back and pulling some of their modalities off the table, right. So as an example, OpenAI, you know, kind of killed Sora, very popular, you know, AI video generator. And I wouldn't be surprised if OpenAI maybe slows down progress on some of their others to really just start focusing on more profitable sectors within the enterprise and just focusing more on their core models. But Gemini is an absolute beast. It does, I'm pretty sure as I'm looking at all my different categories here. Yeah, it does. Every single one. I'm just double checking. Yeah, Image generation, AI, video generation, AI, music generation, design and visual. Yeah, it does, it does everything, Anything here, right. So you can't talk about all these different categories of AI and not absolutely say that Google is dominating this space. Right. Anthropic great models. Cowork Claude. Claude Code some of the most widely used tools over the last three to six months. But Gemini is crushing it across the board in terms of not just multimodality, but in terms of, of having a very powerful tool in every single category. All right, the other thing, Chat GPT is winning the users race, right? Because we did a, in an earlier show on who's winning the AI race. But I think it's important to know as well as you're choosing, oh, you know, what should I be using for AI images? Should I be using, you know, chat, GPT or something else? Well, like we said, you know, OpenAI seems to be cutting back and focusing on some of their services, but there's a lot of other great tools that can lead the way. But still, ChatGPT is absolutely dominating users across the board. And I think understanding the players, even at a base level like I just went over today is extremely important because I think that you have to understand where the platform does enough versus where the specialist wins. Right. Because what can happen is, well, you can start drowning in tools because you can look at all these tools and say, oh wow, wow, I have all these capabilities now that I didn't have before. And depending on what your job is, maybe that's part of your job. Right. But I think some of the best things that the average everyday knowledge worker can do is say no to certain AI tools that would give them new capabilities that they didn't have before. Unless that is a nagging requirement. Right. And it's, it's hard to stay away from that. Right. But I think you have to to stick to where your specialty lies and what category of tools are going to help you expedite that. So here's some advice. You don't, when putting together your ideal, you know, tech stack of the right categories, you don't need every single, you know, oh, I'm going to pick one tool from all these categories. Don't do that like I started with, right? Like I started the show by saying thing, you shouldn't use as many AI tools as I do, you shouldn't be using for the most part, Right. Unless you're a one person company, solopreneur entrepreneur, you know, maybe you're a jack of all trades marketer that has to do way too much. But for the most part, you shouldn't be using one tool from all 11 categories. You know, maybe if you want to learn the basics, you know, set aside a weekend, you know, spend an hour in each category. Sure. But you should just be choosing two to three categories that most align with your daily work. But I will tell you this, you need to be extremely proficient in, well, technically categories 1, 2 and 3 because they're all large language model related. And you at least need to have one tool that you are very good at in category 10, which is the coding copilot, and category 11, which is the general super agents. Because when I talk about this shift from humans doing the work to agents doing the work. I think there's going to be a period of probably, I don't know, 18 to 36 months where if you can commit to those categories or your department or your company, you are going to have an unfair head start on everyone else. I think in 18 to 36 months these things are going to become the de facto way to work work. But we have superhuman AI capabilities that are here now and it's only a matter of time until that just. Well, everyone learns to work that way. But they're here, they're live, they don't require, you know, 30 pounds of duct tape. They're there. All right, so you know, as an example, stick to the tool categories that make the most sense. So you know, a marketer might pair AI search with design tools. An entrepreneur might, you know, pair vibe coding with agents, right? Really stick to the lanes of that. Align with your expertise and what your outputs need to be. But I think the real divide this year is going to be between those who build a real AI stack versus ones who just uses one tool casually versus those who get distracted by a shiny AI syndrome. So you can either be the one that, that just sticks in one category, which is probably not enough. You could be the distracted one, the learner that sticks in every single category. But you're not excelling in any. Or you, you gotta find that middle spot, that sweet spot, right? Knowing the, you know, three to four categories that I mentioned that I think are essential for any knowledge worker and then finding the right tool or maybe one or two additional categories in there that really align with your day to day work. Work. And understanding the full map is where your advantage comes in. Because I think most people are just using one AI category and completely missing the others that are sitting around them. Right? If you're kind of new ish to AI after listening to this, your mind might be completely blown, right? You're like, oh, I didn't know that Google had a tool that could do this. Or I didn't know, you know, Suno could create all this music with your voice. I didn't know Gamma could, you know, take a blog post URL and create a deck out of it with your own image assets. Yes, right. They can do all these things. But you have to ignore shiny object AI syndrome because I think for the most part almost any AI product in the world can fit under one of those 11 parent umbrellas. So pick your categories wisely, the ones that you need to compete in choose your AI operating system or your AI tool of choice and try as much as you can to ignore the others because you need to pick the categories that are going to move your work work forward starting today. All right. I hope this is helpful. I know this is kind of a longer one, but I think it was an important show that we had to do because I get these questions all the time. What tools should I be using? Oh my gosh, like what are AI's capabilities? Well, there you go. Obviously, you know, when it comes to, you know, April, May, June of 2026, I think capabilities are going to change. Maybe these categories are going to change. But right now, if you're listening to this, at least in the first half of 2026, you have a great advantage. You understand all of these AI tools where they fit in the landscape. Now it's up to you to do something about it. I hope this was helpful. If so, please. If you're listening on the podcast, please make sure to subscribe to the show. Leave us a rating if this was helpful and then make sure you go to start here series.com that is going to give you free access to our inner circle community. It's literally not listed anywhere. You can't find this anywhere on the Internet except, well, unless I email it to you. But otherwise, by going to start hereseries.com make sure to go join and check out all of the different episodes in the series that you can go listen to watch for free now. So thank you for tuning in. Hope to see you back tomorrow and every day for more Everyday AI. Thanks y'. All.
