Category: automation

  • Mobile Onboarding A/B testing simply explained

    In earlier posts about Google’s and Twitter’s onboarding tips we mentioned they would absolutely be measuring the impact of Tips and Tours to get the maximum uplift of user understanding and engagement.

    One method is just by simply looking at your analytics and checking the click-thru rate or whatever other CTA (call-to-action) outcome you desired. But 2 big questions loom:

    1. Is what I’m doing going to be a better experience for users?
    2. How do you “continuously improve?

    In recent years – rather than a “spray and pray” approach, it’s favorable to test-and-learn on a subset of your users. Facebook famously run many experiments per day and because their audience size and demographic diversity is massive they can “continuously improve” to killer engagement. If they “burn a few people” along the way its marginal collateral damage in the execution of their bigger goals.

    That sounds mercenary but the “greater-good” is that by learning effectiveness of your experiments will result in better user experiences across the entire user-base and more retained users.

    What do I mean by Mobile Onboarding?

    Onboarding is the early phases of a user’s experience with your App. A wise Product Manager recently said to me “on-boarding doesn’t make a product… …but it can break the product”.

    If you are familiar with Dave McClure’s “startup metrics for pirates” – then the goal of Onboarding is to get the user to the “AR” in “AARRR”. To recap:

    • A – Acquisition
    • A – Activation
    • R – Retention
    • R – Referral
    • R – Revenue

    So Onboarding’s “job” is to get a user Activated and Retended or Retentioned (can I make those words up? OK, OK “Retained”).

    Because a user’s attention-span is slightly worse than a goldfish your best shot is to get the user Activated in the 1st visit. Once they are gone, they may forget you and move onto other tasks.

    Yes – but specifically what do you mean by Onboarding?

    Activation is learning how a user gets to the “ah-ha” moment and cognizes your Apps utility into their “problem solving” model. Specific actions on onboarding are:

    • Get them some instant gratification
    • Get them some more instant gratification
    • Trade some gratification for a favour in return
      • User registration
      • Invite a friend
      • Push notification permission
    • Most importantly it is the education and execution of a task in the App that gets the “ah-ha” moment. This is often:
      • Carousels
      • Tips
      • Tours
      • Coachmarks
      • A guided set of tasks

    Progressive (or Feature) Onboarding

    Any App typically has more than one feature. Many retailers, banks, insurers, real-estate, telcos (and others) have Apps that have multiple nuggets of utility built into the App.

    This is because they have a deep, varied relationship with their customers and multiple features all need to be onboarded. We can’t decide what to call this yet – its “feature” driven – but the goal is to progressively deepen a user’s understanding and extracted value from the App.

    So onboarding (and A/B testing) applies to more than the first “activation” stage of the App.

    What is A/B testing?

    A/B testing, or split testing, are simple experiments to determine which option, A or B, produces a better outcome. It observes the effect of changing a single element, such as the presenting a Tip or Tour to educate a user.

    Champion vs Challenger

    When the process of experimentation is ongoing, the process is known as champion/challenger. The current champion is tested against new challengers to continuously improve the outcome. This is how Contextual allows you to run experiments on an ongoing basis so you can continue to improve your Activation.

    A/B Testing Process

    Step 1: Form a hypothesis around a question you would like to test. The “split” above  might be testing an experiment (based on a hypothesis) that running a Tip or Tour will influence a “Success Metric” of “Purchases”.

    The “Success Metric” does not need to be something so obvious, it may be testing the effectiveness of an experiment to alter “times opened in last 7 days” across the sample population.

    Here’s another example teaching a user how to update their profile and add a selfie.

    Step 2: Know you need statistical significance (or confidence). See the section below on this – it’s a bit statistical but in summary the certainty you want that the outcome of your experiment reflects the truth. Do not simply compare absolute numbers unless the two numbers are so different that you can be sure just by looking at them, such as a difference in conversion rate between 20% and 35%.

    Step 3: Collect enough data to test your hypothesis. With more subtle variations under experiment, more data needs to be collected to make an unambiguous distinction of statistical confidence decided in Step 2.

    Step 4: Analyse the data to draw conclusions. Contextual provides you with the comparison of performance for every campaign grouped by the same “success metric”. The chart below shows the:

    • Blue is the Control Group (Champion)
    • Green is your Experiment  (Challenger)
    • The last 30 days history.

    “Contextual automatically captures screen visits and button clicks without you needing to a-priori think about it”

    Iterate

    Step 5: Build from the conclusions to continue further experiment iterations.

    Sometimes this might mean:

    • Declaring a  new “Champion”
    • Refining a new “Challenger”
    • Or scrapping the hypothesis.

    The most impressive results come from having a culture of ongoing experiments. It will take some time but ultimately the Product Manager can recruit others in their team (developers, QA, growth hackers) to propose other experiments.

    Statistical Significance

    Picking the right metric

    Running experiments are only useful if:

    • You selected the correct “Success Metric” to examine. In Contextual we allow you to automatically chart your “Success Metrics” comparisons, but we also allow you to “what-if” other metrics. Contextual:
    • automatically captures screen visits and button clicks without you needing to a-priori think about it.
    • allows you to sync data from your backend systems so you can measure other out-of-band data like purchases or loyalty points etc.

    A/A/B or A/A/B/B Testing

    It has become more common to also duplicate identical running of an experiment to eliminate any question of statistical biasing using the A/B tool. If there is a variation between A–A or B/B is “statistically significant” then the experiment is invalidated and reject the experiment.

    Sample Size and Significance

    If you toss a coin 2 times its a lousy experiment.  There is an awesome Derren Brown “10 heads in a row” show. Here’s the spoiler video! If you remember back to your statistics classes at College/University the “standard error” (not “standard deviation”) of both A and B need to NOT overlap in order to have significance.

    Where T = test group count and C = converts count and 95% range is 1.96, Standard Error is:

    I’ll do a whole separate post on it for the geeks but using a calculator in the product is good enough for mortals 🙂
    UPDATE: The geek post is here!

    A/B testing vs multivariate testing

    A/A/B is a form of multivariate testing. But multivariate testing is a usually a more complicated form of experimentation that tests changes to several elements of a single page or action at the same time. One example would be testing changes to the colour scheme, picture used and the title font of a landing page.

    The main advantage is being able to see how changes in different elements interact with each other. It is easier to determine the most effective combination of elements using multivariate testing. This whole picture view also allows smaller elements to be tested than A/B testing, since these are more likely to be affected by other components.

    However, since testing multiple variables at once splits up the traffic stream, only sites with substantial amounts of daily traffic are able to conduct meaningful multivariate testing within a reasonable time frame. Each combination of variables must be separated out. For example, if you are testing changes to the colour, font and shape of a call to action button at the same time, each with two options, this results in 8 combinations (2 x 2 x 2) that must be tested at the same time.

    Generally, A/B testing is a better option because of its simplicity in design, implementation and analysis

    Summary

    Experiments can be “spray-and-pray” or they can be run with a discipline that provides statistical certaintly. I’m not saying its an essential step and the ONLY metric you want to apply to your App engagement – but as tools become available to make this testing possible you have the foundations to make it part of you culture.

  • Are you listening to user intent?

    Are you trying to break into the music streaming sector? It’s tough to get in with huge investment already wrapped up in it and some massive players dominating the scene.. It’s probably just as competitive as your sector, right? ????

    We’re going to look at one player from the mobile music streaming sector. Meet “Deezer.”

    You may not have heard of them against Spotify, Pandora and the Apple/Google services, but Deezer has been around for awhile now on Desktop, Mobile and TV devices (I have it on a Western Digital HDMI box). Their App is pretty nice and its approach to curated lists is solid.

    Working on Contextual makes us more aware of when Apps do “feature onboarding” in both good and bad ways. One member of our team is an avid Deezer user and pays for the Premium service. Despite being in the “listening” business, the way Deezer’s user experience is organized shows us that it’s not always easy to listen to user feedback.

    What does poor listening look like?

    Deezer prompts on Android for the user to join the family plan like this:

     

     

     

    Fair enough! These reasons look good and we respect that the family plan is the latest “Hot Hot Hot!!!” upsell technique that all these services use. But…this person doesn’t need the family plan and touches “cancel.” That’s okay, you win some, you lose some!

    Except…everytime this user opens the App he gets the same prompt! This has been going on for weeks and makes Deezer look rather clueless about the negative user experience:

    – It seems hard-coded based on the user’s plan
    – It seems to ignore his intent – what were they thinking!???
    – It’s insensitive to his response
    – It’s alienating a faithful paying customer

    So many Apps use their own homebrewed tips and modals, which is cool but they don’t think to tie the UX to analytics or App behavior.

    Why does this happen??

    In a competitive landscape like mobile music streaming, does Deezer really want to alienate a paying customer? Do you? Here is a possible scenario you’ve experienced at your company that explains why the Deezer example can easily happen to even the best of Apps:

    Marketing has been given the objective to drive sales to this new business model, and the Product and Development teams are keen to support this and get this new promotion or feature out FAST and onto people’s phones. The problem is their capability to develop a homebrewed solution is limited because it doesn’t have the underlying maturity to do this in a way that listens user intent. Instead, they end up irritating their brand new users!

    4 ways Deezer could improve

    So this is what we’d recommend as a solution to this problem:

    1. Get smart with audiences
    Obviously, the Deezer user has moved into a new audience segment – from: “family plan prospect,”
    to: “family plan rejected in January 2017 (or X days ago) more than 2 times.”

    All tips, modals and “feature onboarding” should be targeted at specific audiences. Using a scatter-shot approach and continually offering a feature or offer that users do not want is the in-App equivalent of spam.

    2. Triggers
    When a person opens your App, they have a goal, such as Play a song, Book a Taxi, Buy a product.

    The whole reason you are lucky enough to have your App on this person’s phone is because you have a utility they want.

    So…why the hell would you prompt them when they open the App? Deezer goes one step beyond this bad scenario and prompts on re-foregrounding 🙁

    The best time to prompt a user is:

    Contextually – in a way that’s related to actions they’ve just taken, and right AFTER a happy experience. Let the user have their dopamine shot from your awesome App utility, THEN ask them to help you back. Especially when you want to ask for App ratings as well!

    3. Constructive Nagging (interpreting intent) Mobile users are busy so asking once is not enough. We get that… everyone gets that.

    Make sure to track the number of times the users dismisses your prompt. Try a different channel like push notifications or email. With a platform like Contextual, the open REST/JSON API means those other “out-of-band” events can be part of your audience selection.

    But remember to listen and get out of the way!
    Once the user has dismissed the modal and moved to a new customer audience, this means they have moved on. You should too! Platforms should record the analytics of each user’s interaction and remove the modal from the user experience.

    4. Implement  “Smart Listening” with intelligence and action

    Rather than build a homebrewed solution that has no intelligence and cannot adapt to user responses, Apps can now implement smart onboarding of featues. Contextual simplifies the complexity of:

    – Onboarding and Feature Onboarding Metrics (analytics)

    – Intent interpretation

    – Triggers

    – Automation

    – Measurement

    This is a much smarter approach to feature promotion than rushing code into your latest version just to get the job done. It will take awhile for these platforms to mature to do all the things you might want to hard-code. The benefit to you (and your users) is the agility to provide beautiful tips, tours and modals without the complexity and delay in getting them in front of the user.