Measurement That Makes Sense
Monday and Tuesday I was at the ARF's Audience Measurement 6.0, which used to be a symposium but is now a hash tag (#arfam6). I know at least 50 of you were there too, because I saw you. This is one of my very favorite conferences of the year, because at core, I consider myself to be an audience measurement professional (although, like the ARF's David Marans, I might opt for the larger umbrella term "media research," since the discipline -- as well as the scope of the conference -- is really broader than strictly measuring audiences.) I won't attempt to run down the highlights, save to give a shout-out to a couple of colleagues, Joan FitzGerald and Alan Vaughn, who presented their best-paper-award-winning "Impact of Online and Television Advertising on Consumer Behavior."
On Monday, it seemed like the theme of the day was getting GRPs for the Internet, a topic we've explored more than once in this very space. Then on Tuesday, it seemed as if the theme was, moving beyond GRPs in digital. (When they said things move fast in digital, they weren't kidding!)
The Making Measurement Make Sense (MMMS) initiative, spearheaded by the IAB but now sponsored too by the 4As and the ANA, has declared as one of its major intents the desire to develop a unifying metric for digital campaign evaluation; ideally one that can put digital on comparable footing with TV. When I wrote about MMMS in April, I said that I appreciated the opportunity to participate in that industry dialogue.
Earlier this week, the IAB put out the preliminary findings of Bain's work on MMMS. Take a look. This is important stuff, and the progress this initiative has made in articulating a way forward is impressive; Bain did a great job of ingesting input from across the industry and laying out a construct for moving forward.
For now, I want to offer some thoughts on the topics that were highlighted at #ARFAM6, and directly addressed by MMMS.
The MMMS principles call for moving to an audience-based currency that moves beyond gross impressions, and that focuses on "...the respective reach and frequency of such exposures." My colleagues and I at comScore couldn't agree more.
Gross Rating Points are essentially a known commodity if you can count your impressions against the target audience; and, you know the population size of that target. So the question of whether or not we should have GRPs online is essentially moot. We can count impressions served (more on that a little later), and we generally know the target population. Since most of us can perform division and multiply the result by 100, GRPs are thus known a priori.
The question isn't, are GRPs appropriate in digital; rather, the question that emerges is, are GRPs sufficient? The answer is no.
In analyzing about 600 hundred online campaigns at comScore covering 70 brands in 17 countries, we have come to understand that, far more so than in other media, it is essential for digital advertisers to disentangle the GRP into its component parts, reach and frequency. It is possible that the same number of GRPs can be delivered online with wildly divergent mixes of reach and frequency. It is axiomatic in media planning that desired frequency is between 3 and 10 exposures; three are thought to be required in order to register with the consumer, whereas incremental value is minimal after the tenth exposure.
We've seen campaigns where, among all the persons who were exposed, almost two-thirds of them were only reached once or twice, while about 14% were reached 10 or more times. Worse, two-thirds of the impressions in the campaign were served to people in that 10+ bucket (in other words, we've seen 14% of the reach absorb two thirds of the impressions.) Looking just at the GRPs in this case might suggest that the campaign delivered as planned; but clearly, when looking at how the impressions actually distributed among consumers, such a campaign leaves much to be desired. One key driver for the differences between planned and actual delivery is the fact that ad server cookies are being deleted by about 30% of all Internet users, at a rate approaching 5 times per month. As a result, the ad server is often not able to distinguish a computer that has received impressions already versus one that hasn't. This causes reach to be under-delivered and frequency over-delivered relative to plan.
I've written in the past that reach tells us how many, and frequency tells us how much, but that we have to add a third foundational metric: how good? What was the quality of that impression? How do different ad formats compare in terms of impact and effectiveness? How does the environment in which the ad resides affect the impact of the message? In short, what actually transpires between consumer and ad during that exposure? All this is sometimes called engagement.
The very first published principle of MMMS is to move to the standard of a viewable impression. Again, we couldn't agree more. There can be a big difference between the number of impressions the server sends, and the number that actually make their way onto the user's screen, fully loaded and "above the fold." We've observed that at the campaign level, the share of impressions that never load in a viewable section of screen real estate can run as high as 25% or 30%. This will vary based on the nature of the buy; not surprisingly, premium inventory is more likely to be visible, remnant inventory more likely to remain unseen.
Of course, this is just about exposure; moving from exposure to impact, there are ways to measure effectiveness and ROI.
MMMS says, "The industry must identify and define the specific metrics most valuable to brand marketers and define and implement reliable standards for existing metrics." Here too, we agree. But that's another column.
Before getting to those branding metrics, though, you have to get the messaging in front of consumers. To do so, the fundamentals of media math haven't really changed; you need to think about how many of your target audience were reached; how the impressions among that reach were distributed, and whether the impressions were ever actually seen. These three core building blocks of media effectiveness hold true across all media.
What do you think of MMMS? Let me know in the comments section.__________________________________________________________________________________________
Editor's Note: Dying to dissect data? We're looking for more Metrics Insiders -- folks who can write clearly about trends/issues and/or provide guidance to marketers who want to use metrics more effectively. To be considered for a Metrics Insider spot, please include your industry credentials, as well as one or two article ideas. Writing samples are great, too. Please email firstname.lastname@example.org. Thanks!