Beautiful Snowflakes: The Unique Behaviors of Your Best Customers

B
This is part 12 of 50 in 'The Science of PPC' blog series
The Science of PPC is a 50-part series by Andrew Goodman of Page Zero Media with new content published every Thursday. The goal of the series is to expose readers to enough insights, tips, tricks, and examinations of real-world methodologies to walk away with some feeling of mastery and confidence in their own powers of PPC analysis. Learn more

Earlier today, Maya the cat (pictured below, foreground) was watching snowflakes out the window. She’s just a cat, so for her, it’s a mild diversion before her next nap. She’s never heard all that oohing and aahing about how every snowflake is unique. Being a marketer, and not a cat, the snowflakes made me think about a report in Google Analytics.

Figure 1: Not planning to log into GA

Last week, I talked about absorbing real-world feedback from a frequent e-commerce shopper.

There are also great ways to get a feel for individual user journeys (sometimes, they truly are a long, long, long, and winding, winding, winding road) by tapping into advanced reporting. One such report, User Explorer, is available in many Google Analytics accounts. I’ll focus today on the delights of this report.

But does it really make (“scientific”) sense to drill down on segments so small that they couldn’t be any smaller if they tried? If you will, the “atom level” of the trials and tribulations of a single user? (Those trials and tribulations must be the subatomic particles, then. I mean… who would ever think anyone would need to know subatomic particles are a thing, uhhh, right?) Aren’t aggregates better?

We’ll always plan and optimize around aggregate data, of course.

But actually, some of the greatest minds in marketing and analytics believe that aggregates are ho-hum. Said minds love drilling down on small – even tiny – segments, as a way of building insight about behavior. Avinash Kaushik has even said that “all data in aggregate is crap“! I’m sure he was just trying to get our attention.

Watching how individuals shop is not only fascinating and helpful in shaping marketing strategy, it can drive a stake through the heart of egregious myths about typical behaviors.
Watching how individuals shop – how they research, go away, come back via multiple marketing channels, add to cart, ponder some more, go away, come back, eventually buy, and maybe do the same six or seven times over a period of months or years – is not only fascinating and helpful in shaping marketing strategy, it can drive a stake through the heart of egregious myths about typical behaviors.

A particular class of myths revolve around troubleshooting “what’s wrong with that page,” or “what is causing them to abandon the cart.” Looking at a few real-world journeys (especially across multiple businesses, if you are a marketer who has access to other client data) should be enough to demonstrate that indecision is normal when it comes to certain types of purchases. The fairly typical pattern of adding to cart without buying isn’t something a seller can easily stamp out. Of course, UX troubleshooting is its own discipline, and valuable. But it’s not a silver bullet. Research behaviors simply cannot be stamped out entirely. Nor should they be. They’re normal. Consumers have budgets and tastes. People comparison shop. Sometimes, they need to get feedback from family. Sometimes, they interact with your business (to ask a question) before they buy, as opposed to merely reading your website content, email, or social posts.

Water, water, every where, nor any drop to drink [sic]

Seas of data are available to marketers, yet conjecture abounds. “We figure if they haven’t bought within a couple of days, they’re not going to.” True?

Let User Explorer answer that one! Here’s an interesting example to get us started.

In Google Analytics, find the User Explorer report under Audience > . (Obviously, this is subject to change and may vary from account to account. Hey, it’s Google Analytics! Your cheese may be moved at any time without notice.) First, consider setting a date range of moderate length or longer (say, 4 months to a year or more), if available.

You’ll see this amazing list of individual users, anonymized with long encoded numbers. If you drill into a single one, you’re going to see every user session for which that user has engaged your website. It’s like a calendar of that “user’s journey.” The default sort order is in order of user sessions. If you prefer, sort by revenue, or something else.

For one of our clients (setting a date range of the last six months of 2019), the “champ” in terms of user sessions was a person who visited the website over 300 times, ultimately resulting in just over $300 in revenue. Usually, anyone at the top of any list like this is an outlier, so you can sort of discard these particularly anomalous patterns, but the pattern in itself makes you truly wonder. The prospective customer visited the site hundreds of times from organic search (most of them likely searches on the client’s brand keywords), and then by November, had begun visiting (again, hundreds of times) from a link or multiple links associated with email campaigns (or possibly internal personal emails from family – let’s assume I’m not bright enough to tell the difference from this report). Sometimes, the cart was loaded up with 10 items without buying. Sometimes all of this activity occurred in the middle of the night. This type of buyer we will refer to as Insomniac/Ditherer.

Interestingly, though, at least for the User Explorer report, the user’s “Acquisition Date” and Acquisition source/medium is assigned to the “first time the user interacted with your content.” That was in September, from the Google Organic Search channel.

Even with all of that agony, $300+ is probably around the average for lifetime value of a customer for this client, and we may be just getting warmed up. Perhaps as trust is gained and routines are formed, this customer will go on to spend $2,000 over a lifetime. I don’t suppose that all that visiting is costing the client anything. No servers are getting worn out. The only damage being done is to the user’s thumbs (and REM sleep). Annoying, perhaps, but also interesting, eh?

To protect client and user privacy, I haven’t drilled deeper into the detail of the user sessions, but see a snapshot below. In that additional detail (not shown) every significant pageview and operation (such as “viewed cart,” visited x page, etc.) is detailed and time-stamped. Back in October, the user checked out customer reviews (the ones published on the site). It’s all in there.

Figure 2: Sleepless in Sarnia. (Sarnia is an example: not real client location.)

(Unfortunately, it appears that device-switching behavior is not generally part of the User Explorer report.  The “Client ID” is defined as “a unique ID that Analytics assigns to each device from which users engage your content.” So for this person, imagine their entire family doing similar research on all manner of computers. Their six-month session tally could easily top 1,000!)

Subject #2

But let’s move down the list to the 200th most frequent site visitor for this client in H2 2019. This is an individual who visited the website 34 times between July and October. They spent a healthy $1,135 spread over four transactions. The attribution is divided between “direct” (typing in the web address or perhaps coming in from a bookmark) and PPC advertising (“cpc”). The sad thing about this is we’re going to have a very hard time figuring out how much to credit our different marketing channels. Frequent visits from “Paid Search,” leading to some purchase behavior, may well be from the standard PPC advertising on the client’s main brand, which is similar to a brand organic query in that is purely navigational – returning with high purchase intent.

At another point along the way, the customer spent $180 right after viewing a Display ad. Would it have been a remarketing ad, which serves as a reminder to buy to a person we targeted based on certain audience characteristics (presumably, a previous site visit)? Likely.

About four weeks after buying from the Display ad, this customer spent over $300 (the pattern is familiar: make a test purchase, then expand the relationship once satisfied) from one of the many Paid Search visits subsequent to that. Likely scenario: the product arrived, was high quality, and more needs/wants emerged. On the subsequent “trigger pulling day” in early October, this individual looked at 49 pages on the website and did stuff with the cart more than 30 times, before purchasing 6 items for over $300 in total. The user session was 79 minutes long! The meeting we’ll have to account for all this effort in making a purchase may well last that long as well. ☺ In reality, some purchases involve more blood, sweat and tears than others (say, if they’re customized, or if a whole family is making the decisions). So that’s not as weird as it may sound.

Here’s a snippet of what you’d see in the User Behavior report. I’ve left out some of the more interesting “add to cart” product details to protect user and client privacy.

Figure 3: Cha-Ching.

The insight we gain from this report, then, is in stark contrast to what we learn from “story-less” aggregated data. These are epic quests, narratives worthy of Homer. They are real human lives.

‘Lifetime value’ isn’t merely a cold calculation of profit. We’re also talking about the texture of a person’s relationship with the stuff, software, or services they’re going to surround themselves with.
In the corporate world, we like to truncate the concept of “lifetime value” to “LTV,” almost as if we are ashamed of talking about someone’s lifetime in relation to our mundane product or service. But looked at in a certain way, “lifetime” isn’t merely a cold calculation of profit. We’re also talking about the texture of a person’s relationship with the stuff (or software, or services) they’re going to surround themselves with. The decisions they’re making to beautify their homes or to free up more time by having their car detailed. (Hopefully, no one researched car detailing 338 times before pulling the trigger.) What could be more human?

Some marketers focus a lot on average order sizes, trying to figure out how to induce consumers to pile more into the cart, or to become more interested in higher-ticket items. Nothing wrong with that, of course. But you quickly learn that this alone isn’t the sole profit-driver. A heavy profit-driver is high-value customers! Also known as “high lifetime value.” In evaluating many businesses for investment, investors often use ratios such as LTV:CAC, or “lifetime value divided by the cost to acquire a customer.”

Studying how a single individual made several large orders over a period of 6-8 months, totaling, let’s say, $2,500, is a great way to hammer this point home. That individual is putting the $100 average order size and typical lifetime value of $160 person to shame. In terms of their enthusiasm for the products, and the profit that represents for the business, they’re running circles around Agatha Aggregate. Such insights can become actionable, of course, but maybe the first win from having this insight available is to cut down on the unproductive hand-wringing about the majority of customers who may not be particularly loyal. Or worrying about why people don’t fall down at your feet instantly after interacting with your ads or content. Apparently, people aren’t just moist robots.

The limit does exist

We should be taking our marketing data with a grain of salt. Detailed looks at individual ‘big wins’ in terms of happy customers who took awhile to become so may be enough to reassure us that our efforts in various channels are indeed working, even when the formal data looks pretty lukewarm.
From these long and winding roads, trials, and tribulations (dryly known as the “customer journey” or the “buy cycle” in marketing-speak), we also come to realize that our data collection has its limits. When it takes three months for the cash register to ring (and then, it rings big!), how do all your little marketing segments, like keywords, ads in rotation, etc., fare in terms of enjoying full attribution? Not as well as they should. Beyond the possibly advisable strategy of opening up longer latency windows for conversion tracking, we should also be taking our marketing data with a grain of salt. Detailed looks at individual “big wins” in terms of happy customers who took awhile to become so – entering and exiting your orbit in circuitous ways – may be enough to reassure us that our efforts and spends in various channels are indeed working, or have a very strong shot at being profitable, at least. That’s even when the formal data looks pretty lukewarm. And this may apply particularly to smaller companies who are trying to get liftoff without much positive reinforcement from their early marketing data.

Also from the User Explorer report, I recently noticed that one of the biggest, fastest purchases of a luxury furniture piece from a friend’s recently-launched business was from an Instagram interaction. Thank you, User Explorer! Even if one could imagine that the Instagram-based acquisition of over $1,000 in revenue was not the sole touchpoint for this customer, it provided concrete evidence – backed by the logic that aspirational purchasers do follow specialty retailers on Instagram – that a little sizzle sells. Engagement (paid and unpaid) on any social channel or YouTube isn’t something to sneeze at. Small businesses are stretched for time and resources, but having a little shred of concrete evidence – like a larger-than-usual purchase from a medium that appeals to aspirational buyers – helps them answer the question “should we keep working this channel?”

Figure 4: From the showroom to the owner’s home… and 210 Instagram likes. Hey, if you’re going to post a shot of your own living room on social, we at SOPPC say fair game to publish it! P.S. The wood table is the best! P.P.S. Cool it on the hashtags!

Being “data-driven” may often require a deeper dive than you might expect – and the User Explorer report can be the “data vs. data” reality check you might need. Seeing lukewarm aggregate results for certain channels should be no reason to give up on understanding how consumers think. And when people don’t buy, the answer may be: be patient, they’re getting there. Sometimes the heaviest researchers are those who reward you with the greatest lifetime value when all is said and done. Very special snowflakes indeed.

Read Part 13: Steer Clear of Overconfidence to Win at PPC

About the author

Andrew Goodman

Andrew Goodman is Founder & President of Page Zero Media. His accomplishments include writing the first-ever full-length book about Google AdWords, heading up this Google Ads Premier Partner agency, maintaining a string of 48 consecutive speaking engagements at Search Engine Strategies in North America, co-founding a startup called HomeStars, and wearing the dickens out of a lab coat at the SMX Advanced session called Mad Scientists. His active lifestyle requires increasingly elaborate bowls of yogurt. He works from the Toronto office as well as a home office in Fredericton, NB.

Latest Entries

Tags

About the Author

Andrew Goodman is Founder & President of Page Zero Media. His accomplishments include writing the first-ever full-length book about Google AdWords, heading up this Google Ads Premier Partner agency, maintaining a string of 48 consecutive speaking engagements at Search Engine Strategies in North America, co-founding a startup called HomeStars, and wearing the dickens out of a lab coat at the SMX Advanced session called Mad Scientists. His active lifestyle requires increasingly elaborate bowls of yogurt. He works from the Toronto office as well as a home office in Fredericton, NB.

PPC Management Services

Looking to work with our search marketing agency, Page Zero Media? We’d love to hear from you. Just fill out our form to get a quote

🙂