When conversions are up, but engagement is down
May 2019
We recently ran a homepage redesign test on a B2B lead generation site, and found a winner. Lead submissions were up (yay!) but engagement was down (wha?). How can this happen?
(Engagement here is defined as âclicking on a link or buttonâ.)
Letâs put some numbers around it. For simplicityâs sake, say we observed the following conversion rates:
**Original:
**⢠50% engagement
⢠2% conversion (lead submission)
**Variation:
**⢠40% engagement
⢠3% conversion
And because percents can be hard to reason about, letâs turn them into real numbers. For every 1,000 visitors to each experience, we observe:
Original:
⢠500 engaged visitors
⢠20 conversions
Variation:
⢠400 engaged visitors
⢠30 conversions
Since everyone who converts is also engaged, for every 1,000 visitors weâre really talking about:
Original:
⢠500 - 20 = 480 engaged but not converting
Variation:
⢠400 - 30 = 370 engaged but not converting
The Original experience has 480 - 370 = 110 additional âengaged but not converting visitors.â
So we can rephrase the question âHow can conversions be up when engagement is down?â to âWhat are those extra 110 engaged but not converting visitors doing?â
Depending on how concerned we are, this question is answerable with data. We can segment session recordings by test experience, filter out those who converted or bounced, and watch. Or we can use an analytics tool to find all the events fired and pages visited by each group.
Without getting too deep in the weeds, though, what are the possibilities?
The original experience offers some sort of engagement action that the variation lacks. This action is either not correlated with conversion, or negatively correlated with conversion. Maybe itâs video plays, or some sort of popup that must be dismissed. Whatever it is, we shouldnât care about fewer people doing it if more people are signing up without it.
The original experience is just confusing. Those 110 people are unable to find the information that they need on the homepage, so they click a link, try the navigation, maybe do a search. Presumably they do find what theyâre after, but crucially they do not sign up.
Either of these possible explanations should reinforce our decision to call the variation a winner. Beyond that, they underscore the importance of never calling a test based on engagement. If weâd done so in this case, weâd be costing the business a ton of money.
Now that we have a new homepage experience, weâre free to investigate questions about bounce rate, engagement, and how visitors are using the page. But âmore clicksâ will never be our end goal; there are just too many types of clicks - rage clicks, confused clicks, âno idea what this navigation item means but I sure hope it answers my questionâ clicks. If you optimize for those, youâll make the internet a terrible place. And we both know youâre better than that.