B (12:09)
I think looking at the output of a creative performance definitely matters because that's how you will iterate and find the next winners and or the losers. It is important to look at the relative metrics as well. So maybe what he's referring to is more like don't take it as your source of truth because platform is doing the modeling and such. But at the end of the day, the output, the performance of the creative matters and the approach should be more like what can we learn from it regardless of the output? In a way like good or bad, it's all about the learning. Okay, this creative for whatever reason performed better and other creative performed for whatever reason worse. Let's understand why. What is the hook rate, how's the thumbnail? What is the audio, what is the background and this and that. So once we understand and then we start changing one variable at a time, then it becomes A lot more clearer on how we can engineer the success. Right? I mean, that's the whole point of. That's how you create winners from the winners because you have a winning concept and thanks to the interpretation of the output, you're able to create more winners because you look at the relative metrics. That's a guidance for what you should be testing and what you should be not testing anymore because whatever you do in that particular case doesn't work. However, I have seen from the same concept, like the third variation worked, but the first and second failed. It's the same concept. It's just slightly different messaging and such. So that's why you would always test with few variations just to avoid such cases where you don't just jump to conclusion of a whole concept and so on. I think that's also important. So there could be of course false positives and so on. Nothing is perfect. But also, I'm not also saying like, oh, you should trust algorithm blindly and stuff like that. What I'm just saying is we should be able to interpret the results and then drive conclusions to shape the next creative iterations based on that. So that's when you need to interpret the output. And then what matters at the end of the day is your success rate and not the amount of creatives. Because I think LinkedIn nowadays is full of oh, you know what, we just tested 500 creatives. Oh no, we tested 700. How much did you test? Kind of race. That is not so meaningful because those numbers alone don't matter. If your success rate is very low, I would more look at your cost per successful creative that is beating the BAU ADS performance. And it doesn't matter if you test 20 versus 200. Of course, okay, 20 versus 200 might matter, but it doesn't matter. If you're testing a very high amount of creatives with a little success, it matters more. You test less amount of creatives, stronger hypotheses, stronger investment from an analyzing and hypothesizing way with a higher success rate that what matters more than just the total number of creatives that you are testing. I see potential issues when the creative amount is on the higher end because then the mindset is like, hey, look, I just have to test these. Let's just move on. And then having less emphasis on why it worked, why it didn't work, and so on. And then sometimes there's a disconnect because typically a company that is testing 500 creatives a month tend to have a team and usually there are, I Don't want to call it silos, but the guy who is running the UA is not doing the old briefs of the creatives and then analyzing and then giving the briefs and all that. Typically in the smaller startups, yes, but in the bigger teams there is that. And then, so then the UA person's job is just to run them and maybe analyze them. If they don't have the analyst team, then to give the creative, let's say the optimization manager, to put them into briefs and get the next production. So when the quantity is so high, like I think the level of detail and sometimes also the quality of analyzing and shaping the next input is also being jeopardized a bit because of the high quantity, which is not necessarily a win or something to be proud of alone if your success rate is low. Because also then the mindset shifts, like, hey, Look, I have 500 creatives, I can't spend 5k on all of them. Sometimes this gets into, oh, we only spend $100 on these creatives and that's it. Okay, but how did you decide on that threshold? And I think finding that threshold is such an important thing for every account. For example, there's no universal number that I would give. It's very different from business to business, depending on your cpi, depending on your other CPX metrics. But what I would suggest is to look at your cumulative spend and cumulative CPX metrics, whether you're testing on a CPI or a cost per trial or just cost per subscription and such, and see when they stabilize. So because I see a lot of accounts with high spend, they're sometimes just spending 10k, 20k only to realize, oh, this is actually not a good creative. I'm like, oh, really? I could tell it after 500 bucks. Like, look at the data. It was already bad. It was already screaming, like, why did you spend so much? So in order to avoid that inefficiency and scaling it more efficiently, you could easily look at cumulative spend and cumulative performance over time and to see where your data stabilizes. Because we're not going after the statistical significance. Like no one has kind of money for that. But we also don't want to spend just 20, 30 bucks to see, or we don't want to just let the algorithm decide which one to pick. So there the problem is the following. Like, just because algorithm decides to deliver one creative over others doesn't make the other creatives bad creative. I think that's such a misconception or interpreting in a, I would say in a wrong way. Like that Particular creative that got the 90% of the spend was maybe lucky, we call it like lucky 5000 impressions. Because meta does some sort of a decision making around first five to 10,000 impressions. And if one of those 5,000 people who saw the ad didn't engage or whatever signals that Meta didn't receive, then it will not deliver anymore, but it will deliver to other one. We just got maybe 12 clicks and stuff like that. And then I've seen that multiple times where you take out that dominant ad from that ad set and all the other ads like suddenly bloom and you're like, oh, they were also great ads but because of the other better ad or however you want to call it, which is not necessarily better ad as well, because sometimes Meta called the shots way too early. No conversions, nothing. Like it doesn't even give a chance. So in those cases we don't call that creative that got all the spend successful and all the rest bad labeling. I think this is not universal and this is really, I think case by case basis. I've also seen ad sets with 50 ads live. Oh, let's see what algorithm likes and we will run it for a week and then we will see what happens. And then what happens in that week? 90% goes to the same ad. And because they have this rules of thumb, oh, we run it for seven days and at certain, I don't know, daily budget. But then that's useless when the 90% of the budget goes to one creative. That's not a test. You should in those cases, for example, break the rules and like, hey, look, this reached my threshold, this is my decision threshold. Now I want to see what other ads gonna perform for that minimum threshold that you decide based on your cumulative spend and cumulative performance over time. Yeah, so that's why there is no one rule that is valid for all. But I personally try to get spend on the other creatives. If I see a dominant one, I believe that every creative should have an equal chance. So I democratize a bit within the ad set. Right. And because I've seen it countless times that the decision that made by Meta was way too early. Yes, we trust algorithms, but we want to guide the algorithms. When we see 80, 90% of a dominant asset, let it be a creative, let it be a country and such, you're not fulfilling the potential of the other variables. And I think this is important. For example, if you put the US with all the other smaller, let's say European countries, there's a 90% chance that at least 70% of the spend will go to us because of its size and everything. Also you should never do that anyway. Like you should always tier buy like CAC or LTV depending on what you're optimizing towards and ideally without those dominating variables. And if there's a dominating one, then you can always take it out, test it separately, give a chance to the rest. Because it's not meaningful to have, I don't know, let's say like just a couple of thousand dollars spent with a really good cac, really good roas. Well, if you cannot scale that and then you try to scale by increasing that existing campaign's budget, but that country that you wanted to actually scale is not increasing its spend necessarily because if it's getting dominated by the other bigger countries or based on whatever logic depending on your setup. So in those cases then yes, you can experiment with the value base, the rules and stuff like that and that we can talk about later in the testing and the learning side of things.