Toby Jenkins & Adam Franklin - Creating Your Web Strategy SystemBluewire Media
Toby Jenkins & Adam Franklin from Bluewire Media, kick off the Web Strategy Summit 2012 in Brisbane Australia, with a presentation on how to ultimately create your own web strategy system.
How to Talk about AI to Non-analaysts - Stampedecon AI Summit 2017StampedeCon
While artificial intelligence for self-driving cars and virtual assistants gets a lot of the notion of communicating the needs, effectiveness and measurements is complicated when speaking “geek”! The work of an analyst, however, does not just involve conducting data analysis within but communicating, championing and speaking simply when talking to the organization, clients and management.
Your company is all about saving money to make money. But, did you know that your free spreadsheet based expense report is anything but free?
Play Concur's game, Beat the Sheet, and learn 7 ways your spreadsheet based expense report is really costing you more than you think.
Looking for more ways to cut costs?
Check out our latest whitepaper: http://bit.ly/smb_cutcosts
Steal from the Startups: Growth Tactics for Grownups with Kevin Henrikson Search Engine Journal
Agile marketing tactics aren’t just for startups.
Kevin Henrikson, veteran of two startups sold to Yahoo and Microsoft for $550M, will break down his startup marketing strategies which he subsequently scaled for enterprise marketing teams. (Psst: Kevin is also a co-founder in Alpha Brand Media, Search Engine Journal’s parent company!)
Toby Jenkins & Adam Franklin - Creating Your Web Strategy SystemBluewire Media
Toby Jenkins & Adam Franklin from Bluewire Media, kick off the Web Strategy Summit 2012 in Brisbane Australia, with a presentation on how to ultimately create your own web strategy system.
How to Talk about AI to Non-analaysts - Stampedecon AI Summit 2017StampedeCon
While artificial intelligence for self-driving cars and virtual assistants gets a lot of the notion of communicating the needs, effectiveness and measurements is complicated when speaking “geek”! The work of an analyst, however, does not just involve conducting data analysis within but communicating, championing and speaking simply when talking to the organization, clients and management.
Your company is all about saving money to make money. But, did you know that your free spreadsheet based expense report is anything but free?
Play Concur's game, Beat the Sheet, and learn 7 ways your spreadsheet based expense report is really costing you more than you think.
Looking for more ways to cut costs?
Check out our latest whitepaper: http://bit.ly/smb_cutcosts
Steal from the Startups: Growth Tactics for Grownups with Kevin Henrikson Search Engine Journal
Agile marketing tactics aren’t just for startups.
Kevin Henrikson, veteran of two startups sold to Yahoo and Microsoft for $550M, will break down his startup marketing strategies which he subsequently scaled for enterprise marketing teams. (Psst: Kevin is also a co-founder in Alpha Brand Media, Search Engine Journal’s parent company!)
Retail in the cloud - Risk, Opportunity & Visionultraserve
AWS Summit Sydney - presentation by Samuel Yeats, CEO Ultra Serve
Discussing the risks, opportunities for online retailers as well as considering the potential for new cost models.
The problem
Not so much Winner! Not so much Not so much Not so much Not so much Not so much Not so much Not so much Not so much
The real problem
One month Winner! Two months Three months Four months Five months Six months Seven months Eight months Nine months
1 winner per year
QUOTE FROM ELON MUSK BIOGRAPHY “ He would place this urgency that he expected the revenue in ten years to be ten million dollars a day and that every day we were slower to achieve our goals was a day of missing out on that money.
We’re missing out on this
But…why?
TOO MANY IDEAS!
170 ideas in 3 minutes
Try new CTA Try new CTA location Test multiple CTAs Test different button colors Add an “About” section Try a different headline Add an image to the homepage Shorten my form Auto play the video Try modals for forms Test a free trial Try annual billing Try a high anchor price
Er.
We need a process
Find out what’s stopping folks from converting…
…fix it.
Thanks
I’ll be here all night
Step 1: Take these ideas…
Try new CTA Try new CTA location Test multiple CTAs Test different button colors Add an “About” section Try a different headline Add an image to the homepage Shorten my form Auto play the video Try modals for forms Test a free trial Try annual billing Try a high anchor price
…and throw them away.
Step 2: Listen to your customers and prospects
Try new CTA Try new CTA location Test multiple CTAs Test different button colors Add an “About” section Try a different headline Add an image to the homepage Shorten my form Auto play the video Try modals for forms Test a free trial Try annual billing Try a high anchor price These ideas are good. They just might not be for you.
Gut Your gut is good. It just knows too much.
Curse of knowledge
USER TESTING
Send 3 users through your funnel
Send 3 users to your competition
Send 3 users to a HUGE company in your space
PEEK
QUALAROO
CHAT TRANSCRIPTS
SALES PEOPLE AND SUPPORT
Step 3: Use a framework to make decisions
Test summary Test objective Hypothesis Opportunity size Time to test Likely scenarios Next steps What worked? What didn’t work? HTTP://BIT.LY/ABTEST-FRAMEWORK
The problem
Not so much Winner! Not so much Not so much Not so much Not so much Not so much Not so much Not so much Not so much
The problem opportunity
Not so much Winner! Not so much Not so much Not so much Not so much Not so much Not so much Not so much Not so much
Not so much Winner! Not so much Not so much Not so much Winner! Not so much Not so much Not so much Not so much
Not so much Not so much Not so much Winner! Not so much Not so much Winner! Not so much Not so much Winner!
James D kuhn - 10 lessons to become a successful entrepreneur, I now run various successful businesses and can show you what it takes to be an entrepreneurial
success.decisions is important - especially if you want to become a successful entrepreneur one day.
[Elite Camp 2016] Johnathan Dane - The PPC Thermometer - Why You Should Care ...CXL
Managing millions in PPC ad spend allows for some interesting results. Knowing which combination of PPC placements, offers, ads and more that work best, will help you get the right people into the right parts of your conversion funnel. Plateauing your PPC performance is a fairytale, and Johnathan will show you how to keep getting more and more ROI from your PPC efforts.
Growth Hackers Dublin 6th event on November 3rd. Learn actionable growth hacking and marketing tips to implement into your business the next day. Speakers included Alan O'Rouke, Justin Mares and Jules Coleman of Hassle.com.
For Alan O'Rouke's slides: http://www.slideshare.net/AlexandraSashaBlumen/growth-hackers-dublin-6-alan-orourke-slides-2-of-2
[Elite Camp 2016] Annemarie Klaassen & Tom van de Berg - Moving Beyond Testin...CXL
A/B testing is a great way to really learn from user behavior, but this is mostly not the main objective of an optimization program. It needs to improve revenue! If you're only implementing significant results, you're leaving a lot of money on the table. We will explain how we run optimization at the biggest hotel chain in the Netherlands.
Condensed testing syrup - @OptimiseorDie @sydney sep 2011 - 4 years of testin...Craig Sullivan
A summary of my 4 years of A/B and Split testing, with case studies of work, photography guidelines, and key advice on which elements of the page to test for quick wins.
I enjoyed giving this talk at the Online Retailer Conference in Sydney, which is a fine place to visit.
The presentation c.overs a really good 'pizza' analogy for explaining testing to senior management and budget holders. Then covering how you go about discovering good places to test on your site, and what tools will help you get that data.
Lastly, I explore what worked for me in testing, show some examples of how similar our winners are across the globe and then cover some cross channel testing. The last one here is a big growth area and involves optimising contact centres and channels, using the web as a tool. Some interesting work going on here and I show some of ours, as well as new things in the pipeline.
There are some great resources attached, including a list of remote user testing services and the best 'guides' I could find on 'Conversion Rate Optimisation'. Hope you enjoyed the talk and thank you Sydney.
As innovative businesses reduce their time-to-market and shorten release cycles, the need for continuous delivery methods becomes inevitable. Testers are constantly asked to rapidly and reliably deliver comprehensive test results in ever-decreasing or continuous cycles. Meanwhile, testing is butting heads with hard budget, tight time, and limited resource constraints, causing frustration for everyone involved. How can we turn these pains into gains? First, Ingo Phillip shows how to measure test coverage in terms of risk coverage on both the requirement and test case level—and why this matters. Second, Ingo shows how novel test design techniques help us test smarter, not harder, while keeping the entire test case portfolio straightforward. Third, he demonstrates how this effort results in a robust synthetic test data approach with repeatable, stable test execution. Finally, Ingo explains how to enable continuous delivery through continuous testing without drowning in test cases.
Keynote Ton Wesseling at the Web Analytics Wednesday Copenhagen #wawcph at Se...Online Dialogue
Ton was asked to talk about things that get him excited as a web analyst looking at conversion rate optimization. He picked 5 things:
- The real fun part of web analytics is analyzing how user behavior is changing (analyzing experiments), not creating campaign reports...
- Win: inject your website feedback form responses into your analytics and be able to segment behavior based on goals.
- Run your experiments with an automated and free GTM / GA / EXCEL results set-up!
- To make sure business gets it: apply Bayesian statistics on experiment results, don't report on P values, confidence levels etc.
- Bandit algoritms: use www.smartnotifications.com for automated persuasive messaging on your website.
AWS works with many leading VC's in the UK. Join AWS as we share our learnings about venture capital, how it works and provide tips and tricks on how to pitch with the goal to successfully get funding for your startup.
Retail in the cloud - Risk, Opportunity & Visionultraserve
AWS Summit Sydney - presentation by Samuel Yeats, CEO Ultra Serve
Discussing the risks, opportunities for online retailers as well as considering the potential for new cost models.
The problem
Not so much Winner! Not so much Not so much Not so much Not so much Not so much Not so much Not so much Not so much
The real problem
One month Winner! Two months Three months Four months Five months Six months Seven months Eight months Nine months
1 winner per year
QUOTE FROM ELON MUSK BIOGRAPHY “ He would place this urgency that he expected the revenue in ten years to be ten million dollars a day and that every day we were slower to achieve our goals was a day of missing out on that money.
We’re missing out on this
But…why?
TOO MANY IDEAS!
170 ideas in 3 minutes
Try new CTA Try new CTA location Test multiple CTAs Test different button colors Add an “About” section Try a different headline Add an image to the homepage Shorten my form Auto play the video Try modals for forms Test a free trial Try annual billing Try a high anchor price
Er.
We need a process
Find out what’s stopping folks from converting…
…fix it.
Thanks
I’ll be here all night
Step 1: Take these ideas…
Try new CTA Try new CTA location Test multiple CTAs Test different button colors Add an “About” section Try a different headline Add an image to the homepage Shorten my form Auto play the video Try modals for forms Test a free trial Try annual billing Try a high anchor price
…and throw them away.
Step 2: Listen to your customers and prospects
Try new CTA Try new CTA location Test multiple CTAs Test different button colors Add an “About” section Try a different headline Add an image to the homepage Shorten my form Auto play the video Try modals for forms Test a free trial Try annual billing Try a high anchor price These ideas are good. They just might not be for you.
Gut Your gut is good. It just knows too much.
Curse of knowledge
USER TESTING
Send 3 users through your funnel
Send 3 users to your competition
Send 3 users to a HUGE company in your space
PEEK
QUALAROO
CHAT TRANSCRIPTS
SALES PEOPLE AND SUPPORT
Step 3: Use a framework to make decisions
Test summary Test objective Hypothesis Opportunity size Time to test Likely scenarios Next steps What worked? What didn’t work? HTTP://BIT.LY/ABTEST-FRAMEWORK
The problem
Not so much Winner! Not so much Not so much Not so much Not so much Not so much Not so much Not so much Not so much
The problem opportunity
Not so much Winner! Not so much Not so much Not so much Not so much Not so much Not so much Not so much Not so much
Not so much Winner! Not so much Not so much Not so much Winner! Not so much Not so much Not so much Not so much
Not so much Not so much Not so much Winner! Not so much Not so much Winner! Not so much Not so much Winner!
James D kuhn - 10 lessons to become a successful entrepreneur, I now run various successful businesses and can show you what it takes to be an entrepreneurial
success.decisions is important - especially if you want to become a successful entrepreneur one day.
[Elite Camp 2016] Johnathan Dane - The PPC Thermometer - Why You Should Care ...CXL
Managing millions in PPC ad spend allows for some interesting results. Knowing which combination of PPC placements, offers, ads and more that work best, will help you get the right people into the right parts of your conversion funnel. Plateauing your PPC performance is a fairytale, and Johnathan will show you how to keep getting more and more ROI from your PPC efforts.
Growth Hackers Dublin 6th event on November 3rd. Learn actionable growth hacking and marketing tips to implement into your business the next day. Speakers included Alan O'Rouke, Justin Mares and Jules Coleman of Hassle.com.
For Alan O'Rouke's slides: http://www.slideshare.net/AlexandraSashaBlumen/growth-hackers-dublin-6-alan-orourke-slides-2-of-2
[Elite Camp 2016] Annemarie Klaassen & Tom van de Berg - Moving Beyond Testin...CXL
A/B testing is a great way to really learn from user behavior, but this is mostly not the main objective of an optimization program. It needs to improve revenue! If you're only implementing significant results, you're leaving a lot of money on the table. We will explain how we run optimization at the biggest hotel chain in the Netherlands.
Condensed testing syrup - @OptimiseorDie @sydney sep 2011 - 4 years of testin...Craig Sullivan
A summary of my 4 years of A/B and Split testing, with case studies of work, photography guidelines, and key advice on which elements of the page to test for quick wins.
I enjoyed giving this talk at the Online Retailer Conference in Sydney, which is a fine place to visit.
The presentation c.overs a really good 'pizza' analogy for explaining testing to senior management and budget holders. Then covering how you go about discovering good places to test on your site, and what tools will help you get that data.
Lastly, I explore what worked for me in testing, show some examples of how similar our winners are across the globe and then cover some cross channel testing. The last one here is a big growth area and involves optimising contact centres and channels, using the web as a tool. Some interesting work going on here and I show some of ours, as well as new things in the pipeline.
There are some great resources attached, including a list of remote user testing services and the best 'guides' I could find on 'Conversion Rate Optimisation'. Hope you enjoyed the talk and thank you Sydney.
As innovative businesses reduce their time-to-market and shorten release cycles, the need for continuous delivery methods becomes inevitable. Testers are constantly asked to rapidly and reliably deliver comprehensive test results in ever-decreasing or continuous cycles. Meanwhile, testing is butting heads with hard budget, tight time, and limited resource constraints, causing frustration for everyone involved. How can we turn these pains into gains? First, Ingo Phillip shows how to measure test coverage in terms of risk coverage on both the requirement and test case level—and why this matters. Second, Ingo shows how novel test design techniques help us test smarter, not harder, while keeping the entire test case portfolio straightforward. Third, he demonstrates how this effort results in a robust synthetic test data approach with repeatable, stable test execution. Finally, Ingo explains how to enable continuous delivery through continuous testing without drowning in test cases.
Keynote Ton Wesseling at the Web Analytics Wednesday Copenhagen #wawcph at Se...Online Dialogue
Ton was asked to talk about things that get him excited as a web analyst looking at conversion rate optimization. He picked 5 things:
- The real fun part of web analytics is analyzing how user behavior is changing (analyzing experiments), not creating campaign reports...
- Win: inject your website feedback form responses into your analytics and be able to segment behavior based on goals.
- Run your experiments with an automated and free GTM / GA / EXCEL results set-up!
- To make sure business gets it: apply Bayesian statistics on experiment results, don't report on P values, confidence levels etc.
- Bandit algoritms: use www.smartnotifications.com for automated persuasive messaging on your website.
AWS works with many leading VC's in the UK. Join AWS as we share our learnings about venture capital, how it works and provide tips and tricks on how to pitch with the goal to successfully get funding for your startup.
This webcast is for product owners on how to do market testing in an agile fashion. We've created a methodology called a "Market Sprint" that allows you to very quickly discover and prove out go-to-market models for new or pivoting products. It's a great way to validate requirements and buying patterns in the market before having your team write code.
Decision Making masterclass - Radu Atanasiu, MSM Romania, April 2019LauraMSM
The Decision Making masterclass took place on 18 April at Maastricht School of Management Romania and illustrated the tools we use for decision making, decision frames and evaluating them
User Research: The Superpower Behind Experimentation Programs | VWO WebinarsVWO
User research is often a side-lined activity within the experimentation space, but it’s crucial for keeping your testing velocity up and your win rates high.
Many experimentation and optimization teams fail to make this connection due to a lack of time, resources, knowledge or the siloed structure of their organization.
In this talk, Chris will take you through some practical examples of how user research can drive the quantity of ideas, their quality, and originality, which in turn leads to a much more successful overall experimentation program.
Is Growth appropriate for startup of any stage? What must very early stage or pre-product market fit stage startups focus on. Practical tips to acquire the first 1000 customers or potential customers. The right questions to ask to collect qualitative data. About engagement and retention. About Growth Process, Lean Stack/Lean Startup, SCRUM, Growth Mindset, AARRR Funnel, Pirate Metrics, Funnel Optimization.
Keynote Ton Wesseling at Superweek 2020: How an analyst can add value!Ton Wesseling
How an analyst can add value to your A/B-testing program.
What should be done with the A/B-test program?
A: Increase budgets (More a/b-tests (quantity))
B: Increase knowledge (Better a/b-tests (quality))
C: Decrease budgets (Less a/b-tests (quantity))
As an analyst:
You can calculate the answer
You have a big influence on the outcome
Growth hacking: 5 growth tactics that you can try right now!Eveline Smet
Are you a growth hacker? Or do you want to learn the newest growth hacking tools & tactics? Here are 5 growth hacks that you can try today. We'll be talking about how to steal your competitors's strategy, how to build an e-mail list, how to use analytics for data driven growth, famous growth hacks from Airbnb, Spotify, Twitter etc!
Third parties offer many benefits, but they also break sites, open security holes, and are often bad for performance. Kristian Skoeld shares proven methods and techniques on how to manage third parties and break through IT and business silos. You'll leave with a list of tools, checklists, and approaches to help you end emotionally loaded discussions and make data-driven decisions.
In the first of three presentations, Ali talks about leveraging existing sales and marketing resources and amassing them into a viable Web2Print selling machine. He covers the operational changes necessary to facilitate this pivot and talks about how to manage objections all the way up the hierarchy.
Also included are some top tips on identifying and converting Web2Print sales enquiries and how to address customer objections (because they are always valid!). Lastly, Ali talks about why some Web2Print strategies fail, despite good preparation efforts.
Are you running lots of A/B-tests and is your win ratio through the roof? Or do you have too few winners? In both cases you should be wary of the quality and validity of your experiments. In this session Annemarie will give you 10 concrete tips how to check and warrant the validity of your A/B-tests. This will help you run a more successful optimization program.
Presentation for Digital Elite Day 2020
Ben jij al aan het A/B-testen en behaal je daarmee veel successen? Maar heb je het gevoel dat je niet alle cijfers of uitkomsten kunt vertrouwen? Of vind je juist veel te weinig winnaars? In beide gevallen moet je kritisch kijken naar de kwaliteit van jouw uitgevoerde experimenten.
In deze sessie deelt Annemarie aan de hand van concrete voorbeelden, cases en tips hoe je de validiteit van je A/B-testen kunt waarborgen en checken voor een succesvoller CRO programma.
Workshop at MeasureCamp Amsterdam about building a data driven test strategy. Where can you test? What should you test? How do you analyze the results?
CRO is supposed to be really easy. Everyone can set up an A/B-test in the WYSIWYG editors, the testing tool does all the difficult computations for you and it will tell if you have found a winner. It’s child’s play, right? No, you’re wrong! WYSIWYG editors are very error prone (especially with different browsers) and in order to really analyse and interpret A/B-test results correctly you need a basic understanding of statistics.
This presentation will help you understand:
-The importance of Test Power
-How to correctly set up an A/B-test
-How to analyse test results yourself
-The difference between Frequentist and Bayesian statistics
-How to decide to implement a variation
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
30. @AM_Klaassen / @tomvdberg1
Focus on risk assessment
88,4%
A test result is the probability that B outperforms A:
ranging from 0% - 100%
31. @AM_Klaassen / @tomvdberg1
IMPLEMENT B PROBABILITY * EFFECT ON REVENU
Expected risk 11,6% - € 69.791
Expected uplift 88,4% € 213.530
Contribution € 180.733
* Based on 6 months and an average order value of € 175
Make a risk assessment
Hi everybody! First of all thank you so much for having us over!
We are really excited to be here. Before we start with our real presentation. I would like to the story how we actually came here on stage.
TOM
Somewhere in February Peep posted this tweet, where he asked for female speakers.
TOM
Where I replied to…
TOM
So Peep asked for female speakers. How did I end up on stage?
Annemarie decided to do a duo presentation.
We are both working as a Web Analytics and Optimization Expert at Online Dialogue.
AN
Online Dialogue is a conversion optimization agency in the Netherlands. Some of you may actually be familiar with our crazy bosses Ton (who’s a data geek) and Bart (the obviously crazy psychologist). They speak at several conferences all over the world about data and applied psychology.
We work in multidisciplinary teams consisting of analyst, psychologists, ux designers, developers and project leads.
AN
But, what is it that we do then on a day to day basis?
AN
Well, we mainly focus on getting data insights (and that may be derived from webanalytics data, surveys, heat and scrollmaps, usability tests etcetera. Basically any data we can get our hands on) and combine those with consumer psychology and scientific research.
This combination of data and psychology is then used to come up with several hypothesis to test in order to increase conversion.
AN
As you might expect we run lots and lots of experiment.
AN
We do this for a bunch of clients in the Netherlands and also for some pretty cool international clients.
For most we do high velocity testing. Which means , we run multiple tests per week for them.
AN
The purpose of all those A/B-tests is
on the one hand to add direct value in the short term – you want to increase the revenue that is coming in to the website. That’s what clients sign up for. They want more money out of their site.
And on the other hand in the long run to really learn from user behavior. You want to know what it is that triggers the visitors on your website to be able to use those insights to come up with even better optimization efforts in the future.
TOM
Today we want to show you how we optimize the websites for the biggest hotel chain in the Netherlands. It’s called Van der Valk. Maybe you have heard about it, because they also have a big hotel close to Schiphol airport.
We are the optimization partners for 5 year already.
I will share the practical examples we have and Annemarie will share the theory of the different tactics we are using to optimize at Van der Valk.
TOM
They have 39 hotels in the Netherlands (and 1 in Germany). All the websites have the same look and feel but are unique for every hotel. Every hotel has his own color.
TOM
All the websites are maintained by 1 centralized team. They are paid by a percentage of the total turnover.
We are their Analytics & Optimization partner.
TOM
The manager of van der Valk is mainly interested in earning as much money as possible. His team is dependent upon the revenue they get from online. They get a certain percentage of the revenue.
So our focus is on maximizing ROI and revenue.
How do we do this?
Don’t spend money on stuff you should not spend money on
Maximize number of succesfull experiments.
TOM
Because Van der is Valk is not in the business of science, we are in the business of making money.
TOM
Today we want to share 3 tactics we are using to optimize for money at Van der Valk.
TOM
The first tactic is free test software.
We are only using the test software to inject the test code and equally split the variations. We are not using the drag and drop features or to analyse, we do this in the analytics program.
TOM
The free software we use is GTM testing, which is developed by our colleague Jorrin. GTM testing is a simple and free solution to AB test in Google Tag Manager.
All the test we run at Van der Valk are executed with GTM testing.
Our colleague Jorrin wrote the script and is still adding new features.
TOM
GTM testing is easy to set up and it is Free. For other tool, like VWO and Optimizely you will pay at least a few 100 euro’s. And if you have a high traffic website this will be more then 1000 euros a month.
We also created a easy preview mode to switch between the control and the variation. Our clients can quickly check both variations and give us feedback.
AN
The second tactic we use at Valk to grow their revenue is actually a funny one. Because we proposed a change in statistics we use to evaluate A/B-tests.
We have been using Frequentists statistics with the so called z- or t-test to evaluate A/B-tests. This test evaluation method is most commonly used, but there is a big challenge with this type of evaluation.
AN
The most important challenge is that with a t-test an A/B-test can only have 2 outcomes: you either have a winner of no winner.
And the focus is on finding those real winners. You want to take as little risk as possible.
This is not so surprising if you take into account that t-tests have been used in a lot of medical research as well. Of course you don’t want to bring a medicine to the market if you’re not 100% sure that it won’t make people worse of kill them. You don’t want to take any risk whatsoever.
But businesses aren’t run this way. You need to take some risk in order to grow your business. Right?
Now, Tom will show you an A/B-test we have run at van der Valk which highlights this exact challenge.
TOM
In the data we saw that most visitors (around 50%) where exiting the website at this step in the checkout. Most visitors who where actually starting were also finishing it. So the optimization at this page was not in the form, but in how to get more visitors starting the form.
TOM
We performed an A/B-test with the process indicator in the order flow. In the variation we added an extra step (which you already completed (select a room). This psychological tactic is called ‘set completion’. You make visitors feel they are already half-way their order process and will be more likely to finish their booking.
TOM
Based on this test result you would conclude that it’s no winner, that it mustn’t be implemented and that the measured uplift in conversion rate wasn’t enough. So you will see this a loser and move on to another test idea.
However, there seems to be a positive movement (the measured uplift is 4,55%), but it isn’t big enough to recognize as a significant winner. You probably only need a few more conversions.
TOM
So the conclusion there is no significant effect on conversion, but there are significant more visitors on the 3rd step. This is the last step before customers finish there booking. Based on this numbers we would advise to not implement the variation.
Or maybe retest later with higher volumes.
TOM
But why re-test this. That might be a waste of time and money. Shouldn’t it just be implemented?
AN
If Frequentists statistics confronts us with these kind of challenges, what’s the alternative then?
Well as I said earlier, the most common approach to analysing A/B-tests is the t-test (which is based on frequentist statistics).
But, over the last couple of years more and more software packages (like VWO and Google Optimize) are switching to Bayesian statistics.
And that’s not without reason, because using Bayesian statistics makes more sense, since it better suits how businesses are run and I will show you why.
AN
If you take a look at the test result Tom showed you earlier you see that there was a measured uplift in conversion rate of 4,5%, but this just wasn’t enough to declare it as a significant winner.
You might have a hard time explaining this to your manager, because all he sees is the 4,5% measured uplift.
You would have to explain him about significance, p-values and confidence levels why this is not a winner he needs to implement. But that’s hard!
If you use Bayesian statistics however, to evaluate your A/B-test, then there is no difficult statistical terminology involved. There’s no p-value, z-value, confidence levels et cetara. It just shows you the measured uplift and the probability that B is better than A. Easy right?
Everyone can understand this.
Based on the same numbers of the A/B-test we showed you earlier you have a 88,4% chance that B will actually be better than A.
Probably every manager would like these odds.
AN
In short: the biggest advantage of using a Bayesian A/B-test evaluation method is that it doesn’t have a binary outcome like the t-test does.
A test result won’t tell you winner / no winner, but a percentage between 0 and 100% whether the variation performs better than the original.
In this example 88,4%.
The question that remains is: does the chance of an uplift outweigh the chance of loosing money when you would implement the variation?
AN
What you can do with these numbers is make a proper risk assessment.
When Valk decides to implement the new progress bar they have a 11.6% chance of a drop in revenue of almost 70.000 in 6 months time (and an average order value of 175)
But on the other hand, they also have a 88,4% chance that the variation is actually better and brings in over 200.000 euro.
If you multiply 11.6% times the drop in revenue plus 88.4% times the uplift in revenue and you have the total contribution of this test.
This test is thus expected to generate 180.000 euro in 6 months time.
AN
You could argue that every test with a contribution higher than 0 (that’s when the probability is higher than 50%) should be implemented. But then you don’t take into account the cost of testing
As Tom mentioned earlier Van der Valk is paid by a percentage of their turnover. And running an A/b-test program (we don’t work for free) and implementing variations costs money.
In order to compensate this they are satisfied when the expected contribution of a test is at least 150.000 euro in 6 months.
The conclusion based on a Bayesian test evaluation is thus to implement the variation, because it will probably earn them more money! As opposed to with a frequentist test evaluation you wouldn’t and leave money on the table.
AN
Recently we turned this Bayesian Excel calculator into a webtool as well. It’s for everyone free to use.
If you visit this URL you can input your test and business case data and calculate! It will return the chance that B outperforms A and also the contribution of the test to the business.
AN
In conclusion, if you change the evaluation method of A/B-testing to Bayesian you will have a couple of advantages:
Test results are easier to understand, since you no longer have to communicate difficult statistical terminology
The outcome of a test can be seen as a business case question and is therefor better suitable for commercial businesses
You no longer throw away good test ideas that are nearly significant, because they will probably earn you more money
Because of all this you will have a higher implementation rate and thereby more revenue
TOM
The third tactic we use is a change of test method.
We have been optimizing with just pure A/B-tests.
TOM
In an A/B test you randomly assign equal numbers of users to variation A and variation B. And after this period you decide either to implement B or keep A. Either A or B then receives 100% of the traffic.
TOM
During the test the traffic distribution doesn’t change. If a test runs for 2 weeks. The traffic distribution stays 50% / 50%. We call this pure exploration. During the test period you are exploring which variation is performing better.
After the test the winning variation is determined and implemented (100% exploitation).
If variation B is a winner, then you might regret showing 50% of your traffic an inferior option during the 2 weeks testing.
And also the other way around.
The question is whether this winner is a definite winner. Will it stay that way or is it dependent upon seasonality / the weather and so on?
AN
If running A/B-tests may cause regret, what’s the alternative?
I’ll explain what the difference is between an A/B-test and using so-called Bandits.
I will also highlight in which cases using a bandit is preferred to running normal A/B-tests
AN
Before I will explain how bandit testing works, I will first explain why it’s called bandit testing
The name bandit testing comes from the casino. These slot machines are called one armed bandits, because they tend to rob you of your money.
Each of those one armed bandits has a different payout schedule. A gambler will explore different slot machines to find out which slot machine pays out the most. So he will spend most his money in that slot machine.
But he will still occasionally try the other slot machines as well to make sure he’s still spending his money right. So he exploits the one that pays out the most, but keeps exploring others.
AN
The same principle can be applied to A/B-testing. You want to explore which variation performs better, but when you find out you want to exploit this variation as soon as possible.
The distribution between exploring and exploiting may vary over time.
AN
Now that you know what a bandit is, I will tell you how it works.
When a new visitor enters the site, the visitor is assigned to either Explore or Exploit.
When visitors are assigned to the Explore arm, visitors will be randomly assigned to variation A or B. This is the same as with a pure A/B-test.
The bandit starts with 100% exploring, since you don’t have any knowledge about which variation performs better yet.
AN
When you have collected a number of conversions you have information about the conversion rates of A and B. This historical information is used to determine the division between exploring and exploiting.
Visitors who are assigned to Exploit will be served the best performing variation. If B converts better than A, then B will be shown to new visitors.
The chosen percentage between exploring and exploiting is dynamic and is determined by two factors:
The difference in conversion rate between A and B: the bigger the difference the higher the exploitation rate
And the higher the confidence the higher the exploitation rate. This means that in the beginning of the bandit you’re not so confident in the results yet, but by the end of the test you are.
In the end you might end up with 90% exploiting and 10% exploring.
AN
On a time line this might look something like this. You start with a 50/50 split, but over time – if B has a higher conversion rate – the split becomes like 20/80.
This way you will limit the level of regret of showing the less performing variation to a percentage of your traffic. It also means that by the end of the test your overall conversion rate (and thus revenue) is higher than with a normal A/B-test, because more visitors have seen the better variation.
AN
In which cases is it smart to choose a bandit over an A/B-test?
First of all, when you have a limited time frame to optimize:
For example during promotions. You don’t want to run an A/B-test in this short time frame, because when you have the result, it’s worthless. You want to know which variation performs best during the test.
Secondly, when the time frame is endless and you are doing continuous optimization and are frequently adding and removing versions to be tested. This can be very useful when you face high seasonality in sales and behaviour. A winning variation may perform well in the last-minute season, but perform badly during other seasons. A continuous bandit can solve this problem.
And lastly, when you just simply don’t care much about the learnings, but want to optimize for money. You want the regret during testing to be minimized and average conversion rates to be optimized. This is what van der Valk prefers.
TOM
How did we optimize with bandits at the Van der Valk website?
TOM
At the homepage were 50% enters the websites visitors select their data. After this they receive an overview of all the available rooms + the prices.
TOM
We did several rounds of bandits at the most important page at the Van der Valk website. This is the page were visitors choose their room based on the date they selected.
Based on the impact we need to make and the extra revenue we can earn this is the most important page at the Van der Valk website. And in earlier A/B test we saw that specific message with a clear message can work here.
Therefore we choose this page to start with bandits.
TOM
This are the first 2 messages whe choose to run in the bandit.
“You will immediately receive a confirmation” is chosen because the target group at the website is older and more uncertain. To give them more certainty we tell them they will receive a confirmation immediately.
In a survey we did we found that 50% of the visitors is comparing prices with other websites when they are on the Van der Valk website. To show them when the last booking was made, they get the feeling they can’t wait too long
“Last booking at this hotel was 3 minutes ago”
Which sentences work??
TOM
As Annemarie said, in a bandit the traffic allocation differs per day and per variation.
The first 200 transaction we just run the bandit like an A/B test and keep the traffic split 30% per variation. Otherwise the traffic allocation in the bandit will change with every transaction. If the bandit starts to change the traffic allocation after 200 transactions, the conversion rate per variation is more stable.
The control outside the bandit gets the same percentage all the time (10%).
TOM
If we compare the result of the bandit to a normal A/B test, we earned €7K extra in this period. After a couple of days the conversion rate of variation D was higher compared to the other variations and this wasn’t changing anymore.
Therefore we decided to start a new round and add new variations.
TOM
In the second example we started a new bandit with all the traffic equal again. This round was at the same page as round 1.
> Gratis annuleren: minder uncertaintly voor ouderen + wordt bij veel sites aangeboden (en wordt veel vergeleken)
TOM
As Annemarie said, in a bandit the traffic allocation differs per day and per variation.
The first 200 transaction we just run the bandit like an A/B test and keep the traffic split 33% per variation. Otherwise the traffic allocation in the bandit will change with every transaction. If the bandit starts to change the traffic allocation after 200 transactions, the conversion rate per variation is more stable.
The control outside the bandit gets the same percentage all the time (10%).
TOM
If we compare the result of the bandit to a normal A/B test, we earned €20K extra in this period. After a couple of days the conversion rate of variation D was higher compared to the other variations and this wasn’t changing anymore.
Therefore we decided to start a new round and add new variations.
TOM
AN
To summarize: We talked about 3 tactics:
We use free test software to limit the cost of testing
We use Bayesian statistics to evaluate A/B-tests
And we run Bandit tests
AN
All of this to maximize the revenue of van der Valk
AN
To make our client, Van der Valk happy!!
AN
Because again, we are in the business of making money and not in the business of science.