Transcript
A (0:00)
Dinner time. It's where little moments are cherished. With Blue cash preferred, get 6% cash back at US supermarkets and bring everyone together. I did say everyone. Learn more@americanexpress.com Explore BCP terms and cash back cap apply. With Blue Cash preferred, there's nothing like.
B (0:16)
The American Express Platinum card. Find out your welcome offer after you apply, which could be as high as 175,000 points. Learn more and find out your offer@americanexpress.com explorer Platinum Terms apply. Welcome to the podcast. I'm your host, Jaden Schafer. Today, guys, I am super excited because Google has officially rolled out some music making capabilities inside of Gemini. And the reason why I'm excited is because Gemini has an absolutely massive footprint. So I'm excited to see what comes out of this because this is also getting rolled into YouTube and a bunch of other places. And as someone who personally has been making music and also making money from music since I was in college, this is interesting to me. I've played around with a ton of AI, um, music generators. I've actually made a lot of money from them. I'll cover everything in the AI music industry. I've talked to a lot of artists. My wife is a musician. So I have a lot of interesting background and experience in this area. So I'll tell you why I'm excited about what Google has rolled out. Before we do that, almost even more exciting news. I'm super stoked to announce that AI Box, AI, my own startup, we just launched a complete redesign of our entire platform. It rolled out yesterday. If you haven't tried AI Box yet, I would love for you to check it out. You get access to over 50 of the top models that you can chat with all in one place. You can do this for 8.99amonth so you don't have to have $20 subscriptions to, you know, 20 different platforms. But also we have an entire AI app builder that you can explain what kind of tool you want and it will create that. In the past, we got some feedback that it was a little bit complex, a little bit tricky. We have redesigned and streamlined the entire thing. I personally have spent over two weeks painstakingly redesigning every single UI element. You are welcome. So. So if you want to check out what I have been building, go check out AI Box AI. I would love to hear what you have to say about it. I put blood, sweat and tears into it. And I appreciate all of the people that have been listening to the podcast over the years who have invested in the project and helped us make this a reality. So go check it out at AI Box AI. All right, let's talk about what's going down with Google. So Google has just announced that they are rolling in music features into their AI model. It's DeepMind's Lyria 3 model. So they kind of have these Lyria models that are their music generation. What's cool is it's not just like a standalone website. It's right inside of Gemini where they already have, you know, hundreds of millions of people actively using their Gemini app and gemini.google.com on the on desktop. And so you can actually just chat with it directly. Now. I was playing around this morning trying to check this out and it isn't officially rolled out. So it's give it a couple days, give it a little while. They're going to roll it out to everybody, which is exciting, but it's not currently live, but maybe by the listen to this podcast it will be. So I think it's pretty straightforward. Basically what you can do is you can describe a song that you want Gemini to build. You can ask it to be comical, you can make it R and B, you can make it a slow jam, you can make it rock. And. And basically Gemini is going to generate a 30 second track. It will also have lyrics in it and it's going to create cover art by Nano Banana. So what's interesting is you can actually upload a video or a picture and you can have Gemini make a song that matches the mood of that of that image, which, honestly, I think is a. Is a cool idea. So I'll tell you the good, the bad and the ugly about this product first, because I have watched some demos on it and, you know, from some of the things that it's capable of doing, the first thing that I'll say is this. Lyria 3, it is definitely a big upgrade over Google's earlier music models. These music models, by the way, have been, you know, coming out forever. Way before ChatGPT, I was using AI generated music models, although they were terrible back in the day. But I think I still have some albums from when I was in college that I put out of Symphony AI generated music. That was, you know, an interesting way for me to get into symphony music when I don't, you know, play any symphony instruments. So I considered myself an AI Mozart back then, but when I listened to the music today, I was maybe a little bit far off from Mozart. So this is a big upgrade, which I'm excited about. The cool thing that you can do now is that you can basically make music that's a lot more realistic. It has layered compositions. You can adjust a lot of the elements inside of the song, like the style, the tempo, and the vocals. So you have a lot more control over what it sounds like. And kind of its final sound, it's not just like a static output that you get. You can make some adjustments, which I think is great. And I love these features that they're rolling out. What I will also say is that they're extending Lyria 3 beyond Gemini. So the model is being integrated into YouTube's Dream Track feature, which lets creators basically generate AI music for their video, which I personally think is fantastic. As someone that also has done a lot of YouTube video creation in the past, and this was kind of previously limited to creators in the United States, but now this is going to be expanded globally. So I would expect, you know, a majority of YouTube videos are not created inside of America. So this is a massive expansion that they're doing now. What I will say is that Google has a pretty solid line around the limitations of what this model is able to do. Gemini is definitely not meant to. You know, they're not trying to replace specific artists. If there is a prompt that you put in there and you're like, hey, make me a Johnny Cash song or make me a Beyonce song, the. The whole, you know, the system is not going to create an exact replica of that artist. It's. It perhaps will make something that feels in, like, a similar style or mood, but they're very careful not to, you know, be able to actually clone an actual artist. According to Google, they have a whole bunch of filters that are going to check the outputs against existing content. So they're avoiding any direct cop. And this isn't hard for Google. Right. We know that, like, copyright strikes are a very common thing that Google has some pretty solid technology for. So on YouTube, if you make a YouTube video, even with someone else's, like just a copy of someone else's YouTube video, you can get a copyright strike. Someone else's video, you can get a copyright strike. Someone else's audio and especially somebody else's music, you get a copyright strike. So they have a really good. They have a really good technology. And even I was just recently with my wife, we were trying to figure out the name of a song and we were trying to use Shazam on my iPhone and it couldn't figure out the song. And it was super annoying because I was like trying to hum part of the song and so then I was like, oh gee, what else could we do? And I was like googling it and I realized that Google Music, which I happen to have, also has that feature. So I go to the search bar and Google Music hit the hit like the microphone, hum the song and actually pulled up the song right away. And it was kind of a funny, obscure song. But the point being, this is the exact same technology that Google is going to be using for they. They've cracked this basically way better than even Apple. Shazam. Because Apple is Shazam. You basically have to play the literal song. Like if you're in the grocery store, you Shazam. And it can tell you what songs playing on the pa. Google is way better. And you could just like hum the song in my horrible, probably off key humming, right? So and it's able to pick it up. So Google's going to do a great job of this. I have no doubt. I'm excited about this. They have the right guardrails in place to not ripoff artists. And mostly I think they're just getting themselves out of lawsuits, but maybe even creators are going to stay out of lawsuits. So it's kind of a win win for everyone. Every single song that is generated with Lyria 3 is going to include a synth ID watermark. So basically this signals that it was created with AI. You're not going to be able to hear this to just, you know, like a person listening to the music won't be able to tell, but there's just some frequencies in there that you can put this into an AI detector and it'll be like, yes, this was AI generated. Google's also adding some tools inside of Gemini that let users upload a track and then check if it contains synth markers. So it's kind of interesting. It's like one place where you can generate the music but also check other music to see if it was generated there. This is going to be rolling out to basically Gemini users that are 18 and older everywhere in the world. English, German, Spanish, French, Hindi, Japanese, Korean and Portuguese are all going to be supported so you can make songs in other languages, which is kind of cool. I think right now AI music is definitely something that is growing a lot. We have Udio, we have Suno, we have a lot of other, a lot of other platforms that are kind of working on this. YouTube and Spotify are definitely leaning to this a lot. They're signing deals to monetize AI assisted tracks. And I think at the same time there's a lot of AI companies that are Getting lawsuits from the music industry over training data and copyright claims. So Deezer introduced a system to label AI generated songs. They're trying to limit fraudulent streaming on their platform. Spotify came out with a bunch of AI rules. Although they say you can create or use AI in your songs, but they're, you know, they're obviously telling people not to make 100% AI generated music. But like, they're like, you know what if you are being creative with AI and making music and people like to listen to it and there's no fraudulent streaming going on, like, have at it. So I kind of like that approach that Spotify is taking because obviously this is very useful for artists that perhaps create a song, but they don't play violin, but they need a violin solo for a couple seconds in the background and they want to use AI to generate it. Anyone that's like, you have to use a real person, you know, maybe they can't afford it. Maybe they're from, you know. Yeah, maybe they just can't afford it. And so I think this is incredible. It really sparks creativity and lets you create way more. So I see that 100%. So Lyria 3 is now going to be inside Gemini. I think Google right now is really betting that generative music isn't going to be just a novelty. Right. This is going to be kind of a core part of their creative toolkit going forward. That's why they're embedding this into Gemini itself. And I'm going to say with all that being said and all those features out of the way, I personally don't think Lyria 3 is a series is basically any serious person will ever use it for music creation yet. So I think it's getting there. Google will move it in the right direction, but I just tend to think that there's a couple other platforms, like I think even 11 labs has like music Generator and they just seem a little gimmicky. Google's I mean, in particular generates 30 seconds, which is basically useless unless you're making like a reel. And you just need some generic like elevator music in the background basically. Right. Or like a generic beat in the background. It could be like sort of cool, but it's not gonna be used for anything more than a 30 second clip. So I just don't find that useful at all. What I would say is that personally I've used Suno a ton, created full, fleshed out tracks. Suno is an incredible platform that's specifically for music creation and so obviously I'm gonna find A lot more useful. But they have an entire AI studio where you can sing a song or you could play a song to it. It will create an entire background track. And I have a close friend that has done this. They're a musician. They'll literally. So basically they'll sing Suno and maybe play like a piano or an instrument once, one instrument. And they will have Suno recreate and produce an entire background track with drums, with all the instruments they let it to, you know, they tell it to produce. And they will sing over that track and create a full track. They also have a studio inside of Suno that lets you put your track in and you highlight certain parts of your track. So like, maybe you're like, okay, just from, you know, this timestamp to this timestamp in the song, I want a violin to come in the background and I want percussion to come in the second half. You just select the part of the song and you type in the instrument that you want and it will generate multiple stems below it that you can turn on and off if you like that Are those instruments. It's like, honestly amazing. And if you're a musician that is trying to, you know, you're on a budget in the past. My wife is a musician and she'll have to create her song. She literally had to teach herself banjo because she wrote a song that had the banjo in it and it definitely slowed down her music. It definitely slowed down her music creation on an album that she was writing when she had to spend, you know, a few days learning the banjo for one particular song on the full album. So. And then alternatively, what we also did was on certain songs, we would just go to Fiverr and we'd pay people to play cello or violin in the background, which is fine because we could afford it, but I think a lot of people couldn't afford. Can't afford that. It can get expensive. And to be honest, we got like the. Even the cello person, I think. I think two songs she used it and a third song that she paid to have all of the cello and violin. She just didn't like how it came out. And so we scrapped it. And I don't. Basically wasted the money that we spent on that. So if you were using AI though, we could have just regenerated it 10 times until we found a version that sounded good, that sounded like it fit the mood and the style of the song. It's not like you have one shot for the. For the artist or for a musician to. To give you a track that they created and it wasn't quite your vision, and then you're out the money and you're not bothering them. So anyways, for like a hundred reasons, I think that Suno is incredible for music generation. I have a friend that's a producer. He makes music for Apple, Disney, you know, Chase Bank, a lot of huge organizations for commercials, and he is actively using Suno to create a ton of the music in the background. It sounds incredible. It's amazing. And so I'm excited about Lyria 3 from Google, and I think they're moving in the right direction. I don't think it's quite there yet. I mean, they don't have the, you know, dedicated music studio. So I think there's a lot of these tools that are just dedicated to music and they're doing really good. And then the other thing, from a user perspective, I'm sure a lot of people be mad at me about this, whatever. But from a user perspective, companies like Yurio and Suno, who are both getting sued by, you know, record labels in the music industry for, I think basically using training data from other people's music to train their models. Like, I'm sure a lot of people could be theoretically mad and say that that's wrong, but because of that, their models are way better. Like, I'm just gonna be honest. They. They trained their models on the music, all the music in the world, and their models are way better at making music. Google had, you know, obviously is a responsible company, and so they had a much more limited, like, amount of music to. To pull off of. And so they, their. Their model just isn't as good. So I'm sorry to say that it sounds horrible. And for anyone getting mad at me, I'm sure all of you, many of you love Claude, and guess what? Claude did the exact same thing. They pirated every single book in the entire world. And that's how their model jump started ChatGPT and having a better tone. In the early days of AI, they got a slap on the wrist. Well, not a slap on the wrist. I mean, they had to pay over a billion dollars to settle that to all of the authors. And they also had to physically purchase every single book in the world. Basically, they had a robot that ripped the front covers off and took a picture of every single page and uploaded it to a giant data set which the, you know, basically was found that that's. That is a legal way to train AI models if you buy the actual book, but pirating it was not. So, I mean, I Guess that comes down to like, did Suno purchase a copy of every single song in the world or do they just pirate it? I'm not going to speculate on their training data and anything else, but they are getting sued and their model is way better. So I know that is like, doesn't make a lot of people happy to hear, but as a user, that's basically where I would go and I think where a lot of people will go for creating music. So maybe Google finds a better way to do this and that's a beautiful revolution and artists get compensated. I'll also say there's a lot of. I think things are changing and even Suno is going to have to change up and, and fix some of those mistakes in the future. One thing that I've noticed a lot of organizations are doing, one music distributor in particular I use called Lander, has a banner at the top of your account when you log in that now says opt in to get compensated for AI training. And so you're, you know, as an artist, you can agree to have your music used by AI models to train and you get some sort of compensation and payment for that. So I think we'll see a lot of that in the future and we'll sort out a lot of these kind of problems and lawsuits and eventually, I mean, every artist is going to be many, many artists will be using AI to help assist in their music creation process in one way or another. And I think that they will opt in for monetization. I think all the models will just kind of get that extra library. But for the time being, it seems like Suno and Yu do kind of have a head start due to their training data. All right, thank you so much for tuning into the podcast today. If you want to try out a lot of these AI models that I talk about on the show, like 11 labs for audio creation, go check out AI box AI. There's over 50 AI models on there for 8.99amonth. You get access to all of the top AI models in one place without a million subscriptions or logins. And you can vibe create any tool or workflow that you can imagine. So go check it out, AI Box AI and I will see you guys in the next episode.
