Monday and Tuesday I was at the ARF's Audience Measurement 6.0, which used to be a symposium but is now a hash tag (#arfam6). I know at least 50 of you were there too, because I saw you. This
is one of my very favorite conferences of the year, because at core, I consider myself to be an audience measurement professional (although, like the ARF's David Marans, I might opt for the larger
umbrella term "media research," since the discipline -- as well as the scope of the conference -- is really broader than strictly measuring audiences.) I won't attempt to run down
the highlights, save to give a shout-out to a couple of colleagues, Joan FitzGerald and Alan Vaughn, who presented their best-paper-award-winning "Impact of Online and Television Advertising on
Consumer Behavior."
On Monday, it seemed like the theme of the day was getting GRPs for the Internet, a topic we've explored more than once in this very space. Then on Tuesday, it seemed as if the theme was, moving beyond GRPs in digital.
(When they said things move fast in digital, they weren't kidding!)
The Making Measurement Make Sense (MMMS) initiative, spearheaded by the IAB but now sponsored too by the 4As and
the ANA, has declared as one of its major intents the desire to develop a unifying metric for digital campaign evaluation; ideally one that can put digital on comparable footing with TV. When I
wrote about MMMS in April, I said that I appreciated the opportunity to participate in that industry
dialogue.
Earlier this week, the IAB put out the preliminary
findings of Bain's work on MMMS. Take a look. This is important stuff, and the progress this initiative has made in articulating a way forward is impressive; Bain did a great
job of ingesting input from across the industry and laying out a construct for moving forward.
For now, I want to offer some thoughts on the topics that were highlighted at #ARFAM6, and
directly addressed by MMMS.
The MMMS principles call for moving to an audience-based currency that moves beyond gross impressions, and that focuses on "...the respective reach and
frequency of such exposures." My colleagues and I at comScore couldn't agree more.
Gross Rating Points are essentially a known commodity if you can count your impressions
against the target audience; and, you know the population size of that target. So the question of whether or not we should have GRPs online is essentially moot. We can count impressions
served (more on that a little later), and we generally know the target population. Since most of us can perform division and multiply the result by 100, GRPs are thus known a priori.
The
question isn't, are GRPs appropriate in digital; rather, the question that emerges is, are GRPs sufficient? The answer is no.
In analyzing about 600 hundred online campaigns at
comScore covering 70 brands in 17 countries, we have come to understand that, far more so than in other media, it is essential for digital advertisers to disentangle the GRP into its component parts,
reach and frequency. It is possible that the same number of GRPs can be delivered online with wildly divergent mixes of reach and frequency. It is axiomatic in media planning that desired
frequency is between 3 and 10 exposures; three are thought to be required in order to register with the consumer, whereas incremental value is minimal after the tenth exposure.
We've
seen campaigns where, among all the persons who were exposed, almost two-thirds of them were only reached once or twice, while about 14% were reached 10 or more times. Worse, two-thirds of the
impressions in the campaign were served to people in that 10+ bucket (in other words, we've seen 14% of the reach absorb two thirds of the impressions.) Looking just at the GRPs in this case might
suggest that the campaign delivered as planned; but clearly, when looking at how the impressions actually distributed among consumers, such a campaign leaves much to be desired. One key driver for the
differences between planned and actual delivery is the fact that ad server cookies are being deleted by about 30% of all Internet users, at a rate approaching 5 times per month. As a result, the ad
server is often not able to distinguish a computer that has received impressions already versus one that hasn't. This causes reach to be under-delivered and frequency over-delivered relative to
plan.
I've written in the past that reach tells us how many, and frequency tells us how much, but that we have to add a third foundational metric: how good? What was the
quality of that impression? How do different ad formats compare in terms of impact and effectiveness? How does the environment in which the ad resides affect the impact of the message? In short,
what actually transpires between consumer and ad during that exposure? All this is sometimes called engagement.
The very first published principle of MMMS is to move to the standard of a
viewable impression. Again, we couldn't agree more. There can be a big difference between the number of impressions the server sends, and the number that actually make their way onto
the user's screen, fully loaded and "above the fold." We've observed that at the campaign level, the share of impressions that never load in a viewable section of screen real
estate can run as high as 25% or 30%. This will vary based on the nature of the buy; not surprisingly, premium inventory is more likely to be visible, remnant inventory more likely to remain
unseen.
Of course, this is just about exposure; moving from exposure to impact, there are ways to measure effectiveness and ROI.
MMMS says, "The industry must
identify and define the specific metrics most valuable to brand marketers and define and implement reliable standards for existing metrics." Here too, we agree. But that's another
column.
Before getting to those branding metrics, though, you have to get the messaging in front of consumers. To do so, the fundamentals of media math haven't really changed; you
need to think about how many of your target audience were reached; how the impressions among that reach were distributed, and whether the impressions were ever actually seen. These three core
building blocks of media effectiveness hold true across all media.
What do you think of MMMS? Let me know in the comments section.
__________________________________________________________________________________________
Editor's Note: Dying to dissect data? We're looking for
more Metrics Insiders -- folks who can write clearly about trends/issues and/or provide guidance to marketers who want to use metrics more effectively. To be considered for a Metrics Insider spot,
please include your industry credentials, as well as one or two article ideas. Writing samples are great, too. Please email pfine@mediapost.com. Thanks!