Kontent.ai
Copyright © 2023 Kontent.ai. All rights reserved.
  • Web
  • Privacy policy
  • Cookies policy
  • Consent settings
  • Security
  • GDPR
  • Documentation
  • API reference
  • Product updates
Kontent.ai Learn
  • Plan
  • Set up
  • Model
  • Develop
  • Create

Make your user experience perfect with A/B testing

Is this page helpful?
Complete and continue
  • Sign in
  • David Klement
    9 minutes
    Whatever content you have, it’s important to determine what works with your audience. Whether you want your customers to buy your products or take any other action, you want to tweak your content so that it leads your audience to your goal. A/B testing, also called split testing, is the way to find the path to perfection.

    What is A/B testing

    A/B testing is a randomized experiment in which you test two or more variants of your content against each other. The variants differ in some aspects. You’re comparing which variant gets you closer to your goal. The variants are split among your audience, either evenly, or based on weights given to each variant. Splitting is usually done evenly, but uneven distribution may come in handy if you want to limit the number of visitors who see the newly tested version. That may be the case if the new version contains something about which you don't want to go public in full yet, for instance.

    Do the testing right

    Here are some tips to help you get the most out of your testing. These pieces of advice are valid regardless of the platform you (plan to) test your content on.

    1. Define your goal

    The first thing is to define a goal you want to reach. What do you want your audience to do? Purchase your product? Sign up for a newsletter? Spend more time on your website? And what success rate do you need?

    2. Identify the problem and make a hypothesis

    Now that you have your goal, identify the problem – what is it that keeps you from reaching that goal? Analyze the data about your visitors and develop a hypothesis – what is it that prevents them from taking action? Narrow down the possible causes, so that you don't need to test every single element in your app.

    3. Test the hypothesis

    Once you have the possible causes and their possible solutions, test them using A/B tests. Make sure to always test only one small change. That way, you don't get lost in what was it exactly that lead to a change in your users' behavior.
    For any testing of this kind, you need a big-enough audience. Testing a change on 50 visitors won't give you definitive results because there's a high chance of a coincidence. The optimum is 1000+ visitors per variant for the results to be statistically valid. After the test is finished, re-test the change after some time to eliminate false positives. This is important because the typical pain point of A/B testing is that the test usually lasts a fairly short time.  Re-testing helps you make sure that the change in your customers' behavior wasn't a time-limited fluke caused by, say, upcoming Christmas. If the same test yields similar results repeatedly, the results are way more trust-worthy.

    4. Evaluate the results

    When your testing is done, look at the results and see if any of the tested variants is getting you closer to your goal. If it is, great! Implement it, and optionally start testing another hypothesis to improve your app further. Remember to always look at the results in the context of your whole business. For example, one change may skyrocket the number of individual purchases, but did your total revenue increase correspondingly? You may now have a larger number of smaller purchases which, when summed up, total in a smaller revenue overall.
    If the hypothesis didn't prove right, test another one. A/B testing is an iterative process of trial and error, one false step doesn't mean you should stop trying. 🙂

    How we A/B test in Kontent.ai

    Now that we have the theory covered, let's see how we, at Kontent.ai, do A/B tests. We cover the approach that's proven successful for us and our website here. We have our content highly structured using components and linked items. This makes it easy to select one for testing and leave other parts of the page untouched.

    The content type for A/B tests

    For the purpose of A/B testing, we have one dedicated content type. In content items of this type, we link the original content item we're doing the test for, define the test's time frame, fill in the Google Optimize test ID, and link the variants of the tested item. Each tested variant has its own weight, ID to distinguish it from other variants, and the tested content created according to our hypothesis. The tested content is a linked item of the same content type as the respective component or linked item in the original page. When we serve the A/B-tested page to our visitors, a selected audience gets the page with the alternative tested content instead of the original.

    Our decision-making logic

    When people visit the page we test, we use our own logic to determine which variant of the page to serve them. This decision is based on several criteria:
    • The weights of the test variants determine the ratio of who gets the alternative versions and who gets the original.
    • We look into the visitor’s browser cache to see if there isn’t a cookie that would indicate the user has already visited the page during the test period. If we find it, we serve the variant the user has already got once. This is to ensure that the same person sees the same content consistently.
    • We also check the test's time frame to avoid serving the alternative versions after the test has finished.

    Test results interpretation

    Tracking of user behavior is handled by Google Optimize. When the test is over, we analyze the data to see whether our testing variant has the desired effect. If it did, we implement it. If not, we repeat the test with different changes.
    What's next?
    Manage SEO in JavaScript applications
    When developing any single-page application or a website that uses JavaScript to fetch content from APIs, you need to think carefully about search engine optimization.

    See the diagram on https://viewer.diagrams.net?lightbox=1&nav=1#Uhttps%3A%2F%2Fraw.githubusercontent.com%2FKontent-ai-Learn%2Fkontent-ai-learn-diagrams%2Fmaster%2Fillustrations%2FAB-testing-variants.drawio

    • Develop with Kontent.ai
    • Hello World
    • Hello Web Spotlight
    • Run a sample app
    • Build apps
    • Decide navigation and URLs
    • Environments
    • Get Developer Certification
    An A/B test to verify whether a change of layout is able to keep visitors longer on the page so that they find more information.
    • Optimize images
    • Optimize performance
    • Optimize SEO
    Apart from A/B testing, there are other similar types of tests, such as split-URL testing and multi-variant testing. Read more about experiment types on Google Optimize Resource Hub.
    The tools we use for A/B testing
    • A dedicated content type to create variants of the tested content item or component
    • Our own logic for splitting audience among the variants
    • Google Optimize to measure the performance of each variant
    Choose the tool that suits you the bestKontent.ai is a headless CMS, which means you can combine it with the A/B-testing tools that best suit your needs. There’s Uniform's integration for Kontent.ai, as well as Optimizely, VWO, Qubit, Adobe Target, SiteSpect, and many more.
    • Use sitemaps
    • Personalize content experience
    • A/B test your content
    • Optimize JavaScript apps for SEO
  • What is A/B testing
  • Do the testing right
  • How we A/B test in Kontent.ai