SlideShare una empresa de Scribd logo
1 de 50
18 simple ways to fuck up your
AB testing
28th March 2014 @OptimiseOrDie
@OptimiseOrDie
• UX and Analytics (1999)
• User Centred Design (2001)
• Agile, Startups, No budget (2003)
• Funnel optimisation (2004)
• Multivariate & A/B (2005)
• Conversion Optimisation (2005)
• Persuasive Copywriting (2006)
• Joined Twitter (2007)
• Lean UX (2008)
• Holistic Optimisation (2009)
Was : Group eBusiness Manager, Belron
Now : Spareroom.co.uk
@OptimiseOrDie
Timeline
Tested stupid ideas, lots
Most AB or MVT tests are bullshit
Discovered AB testing
Triage, Triangulation,
Prioritisation, Maths
Zen Plumbing
AB Test Hype Cycle
Craig’s Cynical Quadrant
Improves
revenue
Improves UX
YesNo
No
Yes Client delighted
(and fires you for another UX agency)
Client fucking
delighted
Client absolutely
fucking furious
Client fires you
(then wins an award for your work)
#1 : You’re doing it in the wrong place
@OptimiseOrDie
#1 : You’re doing it in the wrong place
There are 4 areas a CRO expert always looks at:
1. Inbound attrition (medium, source, landing page,
keyword, intent and many more…)
2. Key conversion points (product, basket, registration)
3. Processes and steps (forms, logins, registration, checkout)
4. Layers of engagement (search, category, product, add)
1. Use visitor flow reports for attrition – very useful.
2. For key conversion points, look at loss rates &
interactions
3. Processes and steps – look at funnels or make your own
4. Layers and engagement – make a ring model
@OptimiseOrDie
Examples – Concept
Bounce
Engage
Outcome
@OptimiseOrDie
Examples – 16-25Railcard.co.uk
Bounce
Login to
Account
Content
Engage
Start
Application
Type and
Details
Eligibility
Photo
Complete
@OptimiseOrDie
Examples – Guide Dogs
Bounce
Content
Engage
Donation
Pathway
Donation
Page
Starts
process
Funnel
steps
Complete
@OptimiseOrDie
Within a layer
Page 1
Page 2
Page 3
Page 4 Page 5
Exit
Deeper
Layer
Email
LikeContact
Wishlist
Micro
Conversions
@OptimiseOrDie
#1 : You’re doing it in the wrong place
• Get to know the flow and loss (leaks) inbound, inside and
through key processes or conversion points.
• Once you know the key steps you’re losing people at and
how much traffic you have – make a money model.
• Let’s say 1,000 people see the page a month. Of those,
20% (200) convert to checkout.
• Estimate the influence your test can bring. How much
money or KPI improvement would a 10% lift in the
checkouts deliver?
• Congratulations – you’ve now built the worlds first IT
plan with a return on investment estimate attached!
• I’ll talk more about prioritising later – but a good real
world analogy for you to use:
@OptimiseOrDie
Think like a
store owner!
If you can’t refurbish
the entire store, which
floors or departments
will you invest in
optimising?
Wherever there is:
• Footfall
• Low return
• Opportunity
@OptimiseOrDie
Insight - Inputs
#FAIL
Competitor
copying
Guessing
Dice rolling
An article
the CEO
read
Competitor
change
Panic
Ego
Opinion
Cherished
notions
Marketing
whims Cosmic rays
Not ‘on
brand’
enough
IT
inflexibility
Internal
company
needs
Some
dumbass
consultant
Shiny
feature
blindness
Knee jerk
reactons
#2 : Your hypothesis is crap!
@OptimiseOrDie
Insight - Inputs
Insight
Segmentation
Surveys
Sales and
Call Centre
Session
Replay
Social
analytics
Customer
contact
Eye tracking
Usability
testing
Forms
analytics
Search
analytics Voice of
Customer
Market
research
A/B and
MVT testing
Big &
unstructured
data
Web
analytics
Competitor
evalsCustomer
services
#2 : These are the inputs you need…
@OptimiseOrDie
#2 : Solutions
• You need multiple tool inputs
– Tool decks are here : www.slideshare.net/sullivac
• Usability testing and User facing teams
– If you’re not doing these properly, you’re hosed
• Session replay tools provide vital input
– Get vital additional customer evidence
• Simple page Analytics don’t cut it
– Invest in your analytics, especially event tracking
• Ego, Opinion, Cherished notions – fill gaps
– Fill these vacuums with insights and data
• Champion the user
– Give them a chair at every meeting
@OptimiseOrDie
We believe that doing [A] for
People [B] will make
outcome [C] happen.
We’ll know this when we
observe data [D] and obtain
feedback [E]. (reverse)
@OptimiseOrDie
#3 : No analytics integration
• Investigating problems with tests
• Segmentation of results
• Tests that fail, flip or move around
• Tests that don’t make sense
• Broken test setups
• What drives the averages you see?
@OptimiseOrDie
18
A B B A
These Danish
porn sites are
so hardcore!
We’re still
waiting for our
AB tests to
finish!
• Use a test length calculator like this one:
• visualwebsiteoptimizer.com/ab-split-test-duration/#4 : The test will finish after you die
#5 : You don’t test for long enough
• The minimum length
– 2 business cycles (cross check)
– Usually a week, 2 weeks, Month
– Always test ‘whole’ not partial cycles
– Be aware of multiple cycles
– Don’t self stop!
– PURCHASE CYCLES – KNOW THEM
• How long after that
– I aim for a minimum 250 outcomes, ideally 350+ for each ‘creative’
– If you test 4 recipes, that’s 1400 outcomes needed
– You should have worked out how long each batch of 350 needs before you start!
– 95% confidence or higher is my aim BUT BIG SECRET -> (p values are unreliable)
– If you segment, you’ll need more data
– It may need a bigger sample if the response rates are similar*
– Use a test length calculator but be aware of BARE MINIMUM TO EXPECT
– Important insider tip – watch the error bars! The +/- stuff – let’s explain
* Stats geeks know I’m glossing over something here. That test time depends on how the two
experiments separate in terms of relative performance as well as how volatile the test response is.
I’ll talk about this when I record this one! This is why testing similar stuff sux. 20
#5 : The tennis court
– Let’s say we want to estimate, on average, what height Roger Federer
and Nadal hit the ball over the net at. So, let’s start the match:
@OptimiseOrDie
First Set Federer 6-4
– We start to collect values
62cm
+/- 2cm
63.5cm
+/- 2cm
@OptimiseOrDie
Second Set – Nadal 7-6
– Nadal starts sending them low over the net
62cm
+/- 1cm
62.5cm
+/- 1cm
@OptimiseOrDie
Final Set Nadal 7-6
– We start to collect values
61.8cm
+/- .3cm
62cm
+/- .3cm
Let’s look at this a different way
62.5cm
+/- 1cm
@OptimiseOrDie
9.1%
± 0.3
9.3%
± 0.3
62.5cm
+/- 1cm
@OptimiseOrDie
9.1%
± 0.5
9.3%
± 0.5
9.1%
± 0.2
9.3%
± 0.2
9.1%
± 0.1
9.3%
± 0.1
Graph is a range, not a line:
9.1 ± 0.3%9.1 ± 0.9%9.1 ± 1.9%
#5 : Summary
• The minimum length:
– 2 business cycles and > purchase cycle as a minimum, regardless of
outcomes. Test for less and you’re cutting.
– 250+, prefer 350+ outcomes in each
– Error bar separation between creatives
– 95%+ confidence (unreliable)
• Pay attention to:
– Time it will take for the number of ‘recipes’ in the test
– The actual footfall to the test – not sitewide numbers
– Test results that don’t separate – makes the test longer
– This is why you need brave tests – to drive difference
– The error bars – the numbers in your AB testing tool are not precise –
they’re fuzzy regions that depend on response and sample size.
– Sudden changes in test performance or response
– Monitor early tests like a chef! @OptimiseOrDie
#6 : You peek and jump to conclusions!
• Ignore the graphs. Don’t draw conclusions. Don’t dance. Calm down.
• Get a feel for the test but don’t do anything yet!
• Remember – in A/B - 50% of returning visitors will see a new shiny website!
• Until your test has had at least 1 business cycle and 250-350 outcomes, don’t
bother even getting excited!
• Watching regularly is good though. You’re looking for anything that looks really
odd – your analytics person should be checking all the figures until you’re satisfied
• All tests move around or show big swings early in the testing cycle. Here is a very
high traffic site – it still takes 10 days to start settling. Lower traffic sites will
stretch this period further.
29
#7 : No QA testing for
the AB test?
#7 - QA Test or Die!
• Over 40% of tests have had QA issues.
• It’s very easy to break or bias the testing
Browser testing www.crossbrowsertesting.com
www.browserstack.com
www.spoon.net
www.cloudtesting.com
www.multibrowserviewer.com
www.saucelabs.com
Mobile devices www.perfectomobile.com
www.deviceanywhere.com
www.mobilexweb.com/emulators
www.opendevicelab.com
@OptimiseOrDie
#7 : What other QA testing should I do?
• Cross Browser Testing
• Testing from several locations (office, home, elsewhere)
• Testing the IP filtering is set up
• Test tags are firing correctly (analytics and the test tool)
• Test as a repeat visitor and check session timeouts
• Cross check figures from 2+ sources
• Monitor closely from launch, recheck, watch
• WATCH FOR BIAS!
@OptimiseOrDie
#8 : Opportunities are not prioritised
Once you have a list of potential test
areas, rank them by opportunity vs.
effort.
The common ranking metrics that I
use include:
•Opportunity (revenue, impact)
•Dev resource
•Time to market
•Risk / Complexity
Make yourself a quadrant diagram
and plot them
#9 : Your cycles are too slow
0 6 12 18
Months
Conversion
@OptimiseOrDie
#9 : Solutions
• Give Priority Boarding for opportunities
– The best seats reserved for metric shifters
• Release more often to close the gap
– More testing resource helps, analytics ‘hawk eye’
• Kaizen – continuous improvement
– Others call it JFDI (just f***ing do it)
• Make changes AS WELL as tests, basically!
– These small things add up
• RUSH Hair booking – Over 100 changes
– No functional changes at all – 37% improvement
• Inbetween product lifecycles?
– The added lift for 10 days work, worth 360k
@OptimiseOrDie
#9 : Make your own cycles
@OptimiseOrDie
#10 : How do I know when it’s ready?
• The hallmarks of a cooked test are:
– It’s done at least 1 or preferably 2+ business and at least one if
not two purchase cycles
– You have at least 250-350 outcomes for each recipe
– It’s not moving around hugely at creative or segment level
performance
– The test results are clear – even if the precise values are not
– The intervals are not overlapping (much)
– If a test is still moving around, you need to investigate
– Always declare on a business cycle boundary – not the middle of
a period (this introduces bias)
– Don’t declare in the middle of a limited time period advertising
campaign (e.g. TV, print, online)
– Always test before and after large marketing campaigns (one
week on, one week off)
37
38
#11 : Your test fails
@OptimiseOrDie
#11: Your test fails
• Learn from the failure! If you can’t learn from the failure, you’ve
designed a crap test.
• Next time you design, imagine all your stuff failing. What would
you do? If you don’t know or you’re not sure, get it changed so
that a negative becomes insightful.
• So : failure itself at a creative or variable level should tell you
something.
• On a failed test, always analyse the segmentation and analytics
• One or more segments will be over and under
• Check for varied performance
• Now add the failure info to your Knowledge Base:
• Look at it carefully – what does the failure tell you? Which
element do you think drove the failure?
• If you know what failed (e.g. making the price bigger) then you
have very useful information
• You turned the handle the wrong way
• Now brainstorm a new test
@OptimiseOrDie
#12 : The test is ‘about the same’
• Analyse the segmentation
• Check the analytics and instrumentation
• One or more segments may be over and under
• They may be cancelling out – the average is a lie
• The segment level performance will help you (beware of
small sample sizes)
• If you genuinely have a test which failed to move any
segments, it’s a crap test – be bolder
• This usually happens when it isn’t bold or brave enough in
shifting away from the original design, particularly on
lower traffic sites
• Get testing again!
@OptimiseOrDie
• There are three reasons it is moving around
– Your sample size (outcomes) is still too small
– The external traffic mix, customers or reaction has
suddenly changed or
– Your inbound marketing driven traffic mix is
completely volatile (very rare)
• Check the sample size
• Check all your marketing activity
• Check the instrumentation
• If no reason, check segmentation
#13 : The test keeps moving around
@OptimiseOrDie
• Something like this can happen:
• Check your sample size. If it’s still small, then expect this until the test settles.
• If the test does genuinely flip – and quite severely – then something has changed with the traffic
mix, the customer base or your advertising. Maybe the PPC budget ran out? Seriously!
• To analyse a flipped test, you’ll need to check your segmented data. This is why you have a split
testing package AND an analytics system.
• The segmented data will help you to identify the source of the shift in response to your test. I rarely
get a flipped one and it’s always something changing on me, without being told. The heartless
bastards.
#14 : The test has flipped on me
• No – and this is why:
– It’s a waste of time
– It’s easier to test and monitor instead
– You are eating into test time
– Also applies to A/A/B/B testing
– A/B/A running at 25%/50%/25% is the best
• Read my post here :
http://bit.ly/WcI9EZ
43
#15 : Should I run an A/A test first
#16 : Nobody feels the test
• You promised a 25% rise in checkouts - you only see 2%
• Traffic, Advertising, Marketing may have changed
• Check they’re using the same precise metrics
• Run a calibration exercise
• I often leave a 5 or 10% stub running in a test
• This tracks old creative once new one goes live
• If conversion is also down for that one, BINGO!
• Remember – the AB test is an estimate – it doesn’t
precisely record future performance
• This is why infrequent testing is bad
• Always be trying a new test instead of basking in the
glory of one you ran 6 months ago. You’re only as good
as your next test.
@OptimiseOrDie
#17 : You forgot about Mobile & Tablet
• If you’re AB testing a responsive site, pay attention
• Content will break differently on many screens
• Know thy users and their devices
• Use bango or google analytics to define a test list
• Make sure you test mobile devices & viewports
• What looks good on your desk may not be for the user
• Harder to design cross device tests
• You’ll need to segment mobile, tablet & desktop response
in the analytics or AB testing package
• Your personal phone is not a device mix
• Ask me about making your device list
• Buy core devices, rent the rest from deviceanywhere.com
@OptimiseOrDie
• Forget MVT or A/B/N tests – run your numbers
• Test things with high impact – don’t be a wuss!
• Use UX, Session Replay to aid insight
• Run a task gap survey (4Q style)
• Run a dropped basket survey (LF style)
• Run a general survey + check social + other sites
• Run sitewide tests that appear on all pages or large clusters
of pages –
• UVPs (“We are a cool brand”), USPs (“Free returns!”), UCPs
(“10% off today”).
• Headers, Footers, Nudge Bars, USP bars, footer changes,
Navigation, Product pages, Delivery info etc.
#18 : Oh shit – no traffic!
• If small volumes, contact customers – reach out.
• If data volumes aren’t there, there are still customers!
• Drive design from levers you can apply – game the system
• Pick clean and simple clusters of change (hypothesis driven)
• Use a goal at an earlier ring stage or funnel step
• Beware of using clickthroughs when attrition is high on the
other side
• Try before and after testing on identical time periods
(measure in analytics model)
• Be careful about small sample sizes (<100 outcomes)
• Are you working automated emails?
• Fix JFDI, performance and UX issues too!
#18 : Oh shit – no traffic
Top F***ups for 2014
1. Testing in the wrong place
2. Your hypothesis inputs are crap
3. No analytics integration
4. Your test will finish after you die
5. You don’t test for long enough
6. You peek before it’s ready
7. No QA for your split test
8. Opportunities are not prioritised
9. Testing cycles are too slow
10. You don’t know when tests are ready
11. Your test fails
12. The test is ‘about the same’
13. Test flips behaviour
14. Test keeps moving around
15. You run an A/A test and waste time
16. Nobody ‘feels’ the test
17. You forgot you were responsive
18. You forgot you had no traffic
@OptimiseOrDie
Is there a way to fix this then?
49
Conversion
Heroes!
@OptimiseOrDie
Email
Twitter
: sullivac@gmail.com
: @OptimiseOrDie
: linkd.in/pvrg14
Slides uploaded to SLIDESHARE.NETSULLIVAC
50

Más contenido relacionado

La actualidad más candente

12 Things to do Before Your Company Dies : Conversion Conference London - Oct...
12 Things to do Before Your Company Dies : Conversion Conference London - Oct...12 Things to do Before Your Company Dies : Conversion Conference London - Oct...
12 Things to do Before Your Company Dies : Conversion Conference London - Oct...Craig Sullivan
 
Why do my AB tests suck? measurecamp
Why do my AB tests suck?   measurecampWhy do my AB tests suck?   measurecamp
Why do my AB tests suck? measurecampCraig Sullivan
 
Web Analytics Wednesday - Session Replay Tools are Vital
Web Analytics Wednesday - Session Replay Tools are VitalWeb Analytics Wednesday - Session Replay Tools are Vital
Web Analytics Wednesday - Session Replay Tools are VitalCraig Sullivan
 
Surviving the AB Testing Hype Cycle - Reaktor Breakpoint 2015
Surviving the AB Testing Hype Cycle - Reaktor Breakpoint 2015Surviving the AB Testing Hype Cycle - Reaktor Breakpoint 2015
Surviving the AB Testing Hype Cycle - Reaktor Breakpoint 2015Craig Sullivan
 
Digital Impact 2014 - Oh Boy These AB tests sure look like Bullshit to me
Digital Impact 2014 - Oh Boy These AB tests sure look like Bullshit to meDigital Impact 2014 - Oh Boy These AB tests sure look like Bullshit to me
Digital Impact 2014 - Oh Boy These AB tests sure look like Bullshit to meCraig Sullivan
 
Cross Device Optimisation - Google Analytics Shortcuts
Cross Device Optimisation - Google Analytics ShortcutsCross Device Optimisation - Google Analytics Shortcuts
Cross Device Optimisation - Google Analytics ShortcutsCraig Sullivan
 
Brighton CRO Meetup #1 - Oh Boy These AB tests Sure Look Like Bullshit to Me
Brighton CRO Meetup #1 - Oh Boy These AB tests Sure Look Like Bullshit to MeBrighton CRO Meetup #1 - Oh Boy These AB tests Sure Look Like Bullshit to Me
Brighton CRO Meetup #1 - Oh Boy These AB tests Sure Look Like Bullshit to MeCraig Sullivan
 
Product design is Poo - And how to fix it!
Product design is Poo - And how to fix it!Product design is Poo - And how to fix it!
Product design is Poo - And how to fix it!Craig Sullivan
 
The Neuromarketing Toolkit - Chinwag Psych - 4 Feb 2014
The Neuromarketing Toolkit - Chinwag Psych - 4 Feb 2014The Neuromarketing Toolkit - Chinwag Psych - 4 Feb 2014
The Neuromarketing Toolkit - Chinwag Psych - 4 Feb 2014Craig Sullivan
 
Mobile presentation - Sydney Online Retailer - 26 Sep 2011
Mobile presentation - Sydney Online Retailer - 26 Sep 2011Mobile presentation - Sydney Online Retailer - 26 Sep 2011
Mobile presentation - Sydney Online Retailer - 26 Sep 2011Craig Sullivan
 
eMetrics London - The AB Testing Hype Cycle
eMetrics London - The AB Testing Hype CycleeMetrics London - The AB Testing Hype Cycle
eMetrics London - The AB Testing Hype CycleCraig Sullivan
 
Craig Sullivan - Oh Boy! These A/B tests look like total bullshit! MKTFEST 2014
Craig Sullivan - Oh Boy! These A/B tests look like total bullshit! MKTFEST 2014Craig Sullivan - Oh Boy! These A/B tests look like total bullshit! MKTFEST 2014
Craig Sullivan - Oh Boy! These A/B tests look like total bullshit! MKTFEST 2014Marketing Festival
 
Product Design is Poo - And we're all going to die
Product Design is Poo - And we're all going to dieProduct Design is Poo - And we're all going to die
Product Design is Poo - And we're all going to dieCraig Sullivan
 
Onboard like a juggernaut - Elite camp 2015
Onboard like a juggernaut - Elite camp 2015Onboard like a juggernaut - Elite camp 2015
Onboard like a juggernaut - Elite camp 2015Conversionista
 
Natural born conversion killers - Conversion Jam
Natural born conversion killers - Conversion JamNatural born conversion killers - Conversion Jam
Natural born conversion killers - Conversion JamCraig Sullivan
 
Why Does My Conversion Rate Suck? Craig Sullivan, Senior Optimisation Consult...
Why Does My Conversion Rate Suck? Craig Sullivan, Senior Optimisation Consult...Why Does My Conversion Rate Suck? Craig Sullivan, Senior Optimisation Consult...
Why Does My Conversion Rate Suck? Craig Sullivan, Senior Optimisation Consult...PRWD
 
SAMPLE SIZE – The indispensable A/B test calculation that you’re not making
SAMPLE SIZE – The indispensable A/B test calculation that you’re not makingSAMPLE SIZE – The indispensable A/B test calculation that you’re not making
SAMPLE SIZE – The indispensable A/B test calculation that you’re not makingZack Notes
 
John lewis lego craig sullivan 1335
John lewis lego craig sullivan  1335John lewis lego craig sullivan  1335
John lewis lego craig sullivan 1335Charlie Lines
 
Confessions of an uber optimiser conversion summit - craig sullivan - v 1.9
Confessions of an uber optimiser   conversion summit - craig sullivan - v 1.9Confessions of an uber optimiser   conversion summit - craig sullivan - v 1.9
Confessions of an uber optimiser conversion summit - craig sullivan - v 1.9Craig Sullivan
 
UXPA UK - Toolkits and Tips for Blending UX, Analytics and CRO
UXPA UK - Toolkits and Tips for Blending UX, Analytics and CROUXPA UK - Toolkits and Tips for Blending UX, Analytics and CRO
UXPA UK - Toolkits and Tips for Blending UX, Analytics and CROCraig Sullivan
 

La actualidad más candente (20)

12 Things to do Before Your Company Dies : Conversion Conference London - Oct...
12 Things to do Before Your Company Dies : Conversion Conference London - Oct...12 Things to do Before Your Company Dies : Conversion Conference London - Oct...
12 Things to do Before Your Company Dies : Conversion Conference London - Oct...
 
Why do my AB tests suck? measurecamp
Why do my AB tests suck?   measurecampWhy do my AB tests suck?   measurecamp
Why do my AB tests suck? measurecamp
 
Web Analytics Wednesday - Session Replay Tools are Vital
Web Analytics Wednesday - Session Replay Tools are VitalWeb Analytics Wednesday - Session Replay Tools are Vital
Web Analytics Wednesday - Session Replay Tools are Vital
 
Surviving the AB Testing Hype Cycle - Reaktor Breakpoint 2015
Surviving the AB Testing Hype Cycle - Reaktor Breakpoint 2015Surviving the AB Testing Hype Cycle - Reaktor Breakpoint 2015
Surviving the AB Testing Hype Cycle - Reaktor Breakpoint 2015
 
Digital Impact 2014 - Oh Boy These AB tests sure look like Bullshit to me
Digital Impact 2014 - Oh Boy These AB tests sure look like Bullshit to meDigital Impact 2014 - Oh Boy These AB tests sure look like Bullshit to me
Digital Impact 2014 - Oh Boy These AB tests sure look like Bullshit to me
 
Cross Device Optimisation - Google Analytics Shortcuts
Cross Device Optimisation - Google Analytics ShortcutsCross Device Optimisation - Google Analytics Shortcuts
Cross Device Optimisation - Google Analytics Shortcuts
 
Brighton CRO Meetup #1 - Oh Boy These AB tests Sure Look Like Bullshit to Me
Brighton CRO Meetup #1 - Oh Boy These AB tests Sure Look Like Bullshit to MeBrighton CRO Meetup #1 - Oh Boy These AB tests Sure Look Like Bullshit to Me
Brighton CRO Meetup #1 - Oh Boy These AB tests Sure Look Like Bullshit to Me
 
Product design is Poo - And how to fix it!
Product design is Poo - And how to fix it!Product design is Poo - And how to fix it!
Product design is Poo - And how to fix it!
 
The Neuromarketing Toolkit - Chinwag Psych - 4 Feb 2014
The Neuromarketing Toolkit - Chinwag Psych - 4 Feb 2014The Neuromarketing Toolkit - Chinwag Psych - 4 Feb 2014
The Neuromarketing Toolkit - Chinwag Psych - 4 Feb 2014
 
Mobile presentation - Sydney Online Retailer - 26 Sep 2011
Mobile presentation - Sydney Online Retailer - 26 Sep 2011Mobile presentation - Sydney Online Retailer - 26 Sep 2011
Mobile presentation - Sydney Online Retailer - 26 Sep 2011
 
eMetrics London - The AB Testing Hype Cycle
eMetrics London - The AB Testing Hype CycleeMetrics London - The AB Testing Hype Cycle
eMetrics London - The AB Testing Hype Cycle
 
Craig Sullivan - Oh Boy! These A/B tests look like total bullshit! MKTFEST 2014
Craig Sullivan - Oh Boy! These A/B tests look like total bullshit! MKTFEST 2014Craig Sullivan - Oh Boy! These A/B tests look like total bullshit! MKTFEST 2014
Craig Sullivan - Oh Boy! These A/B tests look like total bullshit! MKTFEST 2014
 
Product Design is Poo - And we're all going to die
Product Design is Poo - And we're all going to dieProduct Design is Poo - And we're all going to die
Product Design is Poo - And we're all going to die
 
Onboard like a juggernaut - Elite camp 2015
Onboard like a juggernaut - Elite camp 2015Onboard like a juggernaut - Elite camp 2015
Onboard like a juggernaut - Elite camp 2015
 
Natural born conversion killers - Conversion Jam
Natural born conversion killers - Conversion JamNatural born conversion killers - Conversion Jam
Natural born conversion killers - Conversion Jam
 
Why Does My Conversion Rate Suck? Craig Sullivan, Senior Optimisation Consult...
Why Does My Conversion Rate Suck? Craig Sullivan, Senior Optimisation Consult...Why Does My Conversion Rate Suck? Craig Sullivan, Senior Optimisation Consult...
Why Does My Conversion Rate Suck? Craig Sullivan, Senior Optimisation Consult...
 
SAMPLE SIZE – The indispensable A/B test calculation that you’re not making
SAMPLE SIZE – The indispensable A/B test calculation that you’re not makingSAMPLE SIZE – The indispensable A/B test calculation that you’re not making
SAMPLE SIZE – The indispensable A/B test calculation that you’re not making
 
John lewis lego craig sullivan 1335
John lewis lego craig sullivan  1335John lewis lego craig sullivan  1335
John lewis lego craig sullivan 1335
 
Confessions of an uber optimiser conversion summit - craig sullivan - v 1.9
Confessions of an uber optimiser   conversion summit - craig sullivan - v 1.9Confessions of an uber optimiser   conversion summit - craig sullivan - v 1.9
Confessions of an uber optimiser conversion summit - craig sullivan - v 1.9
 
UXPA UK - Toolkits and Tips for Blending UX, Analytics and CRO
UXPA UK - Toolkits and Tips for Blending UX, Analytics and CROUXPA UK - Toolkits and Tips for Blending UX, Analytics and CRO
UXPA UK - Toolkits and Tips for Blending UX, Analytics and CRO
 

Destacado

The Million Dollar Optimization Strategy - Andre Morys - ConversionXL Live 2015
The Million Dollar Optimization Strategy - Andre Morys - ConversionXL Live 2015The Million Dollar Optimization Strategy - Andre Morys - ConversionXL Live 2015
The Million Dollar Optimization Strategy - Andre Morys - ConversionXL Live 2015CXL
 
Conversion rate optimisation 101 @ LS Yorkshire Leeds Feb 2014
Conversion rate optimisation 101 @ LS Yorkshire Leeds Feb 2014Conversion rate optimisation 101 @ LS Yorkshire Leeds Feb 2014
Conversion rate optimisation 101 @ LS Yorkshire Leeds Feb 2014Manuel Da Costa
 
Introduction to conversion rate optimization (cro) for ecommerce
Introduction to conversion rate optimization (cro) for ecommerceIntroduction to conversion rate optimization (cro) for ecommerce
Introduction to conversion rate optimization (cro) for ecommerceIn Marketing We Trust
 
Conversion Rate Optimization 101 - Converting Visitors In To Buyers
Conversion Rate Optimization 101 - Converting Visitors In To BuyersConversion Rate Optimization 101 - Converting Visitors In To Buyers
Conversion Rate Optimization 101 - Converting Visitors In To BuyersKothapally Arun
 
Conversion Strategy (DE)
Conversion Strategy (DE)Conversion Strategy (DE)
Conversion Strategy (DE)Andre Morys
 
Basics of AB testing in online products
Basics of AB testing in online productsBasics of AB testing in online products
Basics of AB testing in online productsAshish Dua
 
Ab testing work flow
Ab testing work flowAb testing work flow
Ab testing work flowJainul Khan
 
Conversion Rate Optimization 101 - Kick-Start Your Growth Engine
Conversion Rate Optimization 101 - Kick-Start Your Growth EngineConversion Rate Optimization 101 - Kick-Start Your Growth Engine
Conversion Rate Optimization 101 - Kick-Start Your Growth EngineKissmetrics on SlideShare
 
Apache Flink at Strata San Jose 2016
Apache Flink at Strata San Jose 2016Apache Flink at Strata San Jose 2016
Apache Flink at Strata San Jose 2016Kostas Tzoumas
 
Continuous Processing with Apache Flink - Strata London 2016
Continuous Processing with Apache Flink - Strata London 2016Continuous Processing with Apache Flink - Strata London 2016
Continuous Processing with Apache Flink - Strata London 2016Stephan Ewen
 

Destacado (11)

The Million Dollar Optimization Strategy - Andre Morys - ConversionXL Live 2015
The Million Dollar Optimization Strategy - Andre Morys - ConversionXL Live 2015The Million Dollar Optimization Strategy - Andre Morys - ConversionXL Live 2015
The Million Dollar Optimization Strategy - Andre Morys - ConversionXL Live 2015
 
Conversion rate optimisation 101 @ LS Yorkshire Leeds Feb 2014
Conversion rate optimisation 101 @ LS Yorkshire Leeds Feb 2014Conversion rate optimisation 101 @ LS Yorkshire Leeds Feb 2014
Conversion rate optimisation 101 @ LS Yorkshire Leeds Feb 2014
 
Introduction to conversion rate optimization (cro) for ecommerce
Introduction to conversion rate optimization (cro) for ecommerceIntroduction to conversion rate optimization (cro) for ecommerce
Introduction to conversion rate optimization (cro) for ecommerce
 
Ab testing
Ab testingAb testing
Ab testing
 
Conversion Rate Optimization 101 - Converting Visitors In To Buyers
Conversion Rate Optimization 101 - Converting Visitors In To BuyersConversion Rate Optimization 101 - Converting Visitors In To Buyers
Conversion Rate Optimization 101 - Converting Visitors In To Buyers
 
Conversion Strategy (DE)
Conversion Strategy (DE)Conversion Strategy (DE)
Conversion Strategy (DE)
 
Basics of AB testing in online products
Basics of AB testing in online productsBasics of AB testing in online products
Basics of AB testing in online products
 
Ab testing work flow
Ab testing work flowAb testing work flow
Ab testing work flow
 
Conversion Rate Optimization 101 - Kick-Start Your Growth Engine
Conversion Rate Optimization 101 - Kick-Start Your Growth EngineConversion Rate Optimization 101 - Kick-Start Your Growth Engine
Conversion Rate Optimization 101 - Kick-Start Your Growth Engine
 
Apache Flink at Strata San Jose 2016
Apache Flink at Strata San Jose 2016Apache Flink at Strata San Jose 2016
Apache Flink at Strata San Jose 2016
 
Continuous Processing with Apache Flink - Strata London 2016
Continuous Processing with Apache Flink - Strata London 2016Continuous Processing with Apache Flink - Strata London 2016
Continuous Processing with Apache Flink - Strata London 2016
 

Similar a #Measurecamp : 18 Simple Ways to F*** up Your AB Testing

AB Testing and UX - a love story with numbers and people (by Craig Sullivan a...
AB Testing and UX - a love story with numbers and people (by Craig Sullivan a...AB Testing and UX - a love story with numbers and people (by Craig Sullivan a...
AB Testing and UX - a love story with numbers and people (by Craig Sullivan a...Northern User Experience
 
20 Ways to Shaft your Split Tesring : Conversion Conference
20 Ways to Shaft your Split Tesring : Conversion Conference20 Ways to Shaft your Split Tesring : Conversion Conference
20 Ways to Shaft your Split Tesring : Conversion ConferenceCraig Sullivan
 
Conversion Hotel 2014: Craig Sullivan (UK) keynote
Conversion Hotel 2014: Craig Sullivan (UK) keynoteConversion Hotel 2014: Craig Sullivan (UK) keynote
Conversion Hotel 2014: Craig Sullivan (UK) keynoteWebanalisten .nl
 
Toolkits and tips for UX analytics CRO by Craig Sullivan
Toolkits and tips for UX analytics CRO by Craig SullivanToolkits and tips for UX analytics CRO by Craig Sullivan
Toolkits and tips for UX analytics CRO by Craig SullivanUXPA UK
 
Anton Muzhailo - Practical Test Process Improvement using ISTQB
Anton Muzhailo - Practical Test Process Improvement using ISTQBAnton Muzhailo - Practical Test Process Improvement using ISTQB
Anton Muzhailo - Practical Test Process Improvement using ISTQBIevgenii Katsan
 
Toolkits and tips of the conversion pros v 1.6
Toolkits and tips of the conversion pros v 1.6Toolkits and tips of the conversion pros v 1.6
Toolkits and tips of the conversion pros v 1.6Craig Sullivan
 
Fail and Win: Why a Failed Test Isn’t a Bad Thing
Fail and Win: Why a Failed Test Isn’t a Bad ThingFail and Win: Why a Failed Test Isn’t a Bad Thing
Fail and Win: Why a Failed Test Isn’t a Bad ThingOptimizely
 
Conversion Optimization Willa Fogarty
Conversion Optimization Willa FogartyConversion Optimization Willa Fogarty
Conversion Optimization Willa FogartyWilla Fogarty
 
Case Study by Jon Warden and Shruti Ahuja of Haymarket - Optimizely Experienc...
Case Study by Jon Warden and Shruti Ahuja of Haymarket - Optimizely Experienc...Case Study by Jon Warden and Shruti Ahuja of Haymarket - Optimizely Experienc...
Case Study by Jon Warden and Shruti Ahuja of Haymarket - Optimizely Experienc...Optimizely
 
Google Analytics Powerups and Smartcuts
Google Analytics Powerups and Smartcuts Google Analytics Powerups and Smartcuts
Google Analytics Powerups and Smartcuts Charles Meaden
 
Patrick McKenzie Opticon 2014: Advanced A/B Testing
Patrick McKenzie Opticon 2014: Advanced A/B TestingPatrick McKenzie Opticon 2014: Advanced A/B Testing
Patrick McKenzie Opticon 2014: Advanced A/B TestingPatrick McKenzie
 
Customer Development Fast Protyping
Customer Development Fast ProtypingCustomer Development Fast Protyping
Customer Development Fast ProtypingSerdar Temiz
 
Growing Up The Right Way: An Example of How to Scale When You're Aren't a Sta...
Growing Up The Right Way: An Example of How to Scale When You're Aren't a Sta...Growing Up The Right Way: An Example of How to Scale When You're Aren't a Sta...
Growing Up The Right Way: An Example of How to Scale When You're Aren't a Sta...Sauce Labs
 
Craig Sullivan - Keynote speaker summary & final thoughts - Conversion Hotel ...
Craig Sullivan - Keynote speaker summary & final thoughts - Conversion Hotel ...Craig Sullivan - Keynote speaker summary & final thoughts - Conversion Hotel ...
Craig Sullivan - Keynote speaker summary & final thoughts - Conversion Hotel ...Webanalisten .nl
 
iSG Webinar – AB Testing: The most important thing you’re NOT doing
iSG Webinar – AB Testing: The most important thing you’re NOT doingiSG Webinar – AB Testing: The most important thing you’re NOT doing
iSG Webinar – AB Testing: The most important thing you’re NOT doingBridgeline Digital
 
Research and Discovery Tools for Experimentation - 17 Apr 2024 - v 2.3 (1).pdf
Research and Discovery Tools for Experimentation - 17 Apr 2024 - v 2.3 (1).pdfResearch and Discovery Tools for Experimentation - 17 Apr 2024 - v 2.3 (1).pdf
Research and Discovery Tools for Experimentation - 17 Apr 2024 - v 2.3 (1).pdfVWO
 
Management Consulting Productivity Hacks
Management Consulting Productivity HacksManagement Consulting Productivity Hacks
Management Consulting Productivity HacksAsen Gyczew
 
OEE Interphex New York 2013
OEE Interphex New York 2013OEE Interphex New York 2013
OEE Interphex New York 2013Adrian Pask
 

Similar a #Measurecamp : 18 Simple Ways to F*** up Your AB Testing (20)

AB Testing and UX - a love story with numbers and people (by Craig Sullivan a...
AB Testing and UX - a love story with numbers and people (by Craig Sullivan a...AB Testing and UX - a love story with numbers and people (by Craig Sullivan a...
AB Testing and UX - a love story with numbers and people (by Craig Sullivan a...
 
20 Ways to Shaft your Split Tesring : Conversion Conference
20 Ways to Shaft your Split Tesring : Conversion Conference20 Ways to Shaft your Split Tesring : Conversion Conference
20 Ways to Shaft your Split Tesring : Conversion Conference
 
Conversion Hotel 2014: Craig Sullivan (UK) keynote
Conversion Hotel 2014: Craig Sullivan (UK) keynoteConversion Hotel 2014: Craig Sullivan (UK) keynote
Conversion Hotel 2014: Craig Sullivan (UK) keynote
 
Toolkits and tips for UX analytics CRO by Craig Sullivan
Toolkits and tips for UX analytics CRO by Craig SullivanToolkits and tips for UX analytics CRO by Craig Sullivan
Toolkits and tips for UX analytics CRO by Craig Sullivan
 
Anton Muzhailo - Practical Test Process Improvement using ISTQB
Anton Muzhailo - Practical Test Process Improvement using ISTQBAnton Muzhailo - Practical Test Process Improvement using ISTQB
Anton Muzhailo - Practical Test Process Improvement using ISTQB
 
Toolkits and tips of the conversion pros v 1.6
Toolkits and tips of the conversion pros v 1.6Toolkits and tips of the conversion pros v 1.6
Toolkits and tips of the conversion pros v 1.6
 
Fail and Win: Why a Failed Test Isn’t a Bad Thing
Fail and Win: Why a Failed Test Isn’t a Bad ThingFail and Win: Why a Failed Test Isn’t a Bad Thing
Fail and Win: Why a Failed Test Isn’t a Bad Thing
 
The Finishing Line
The Finishing LineThe Finishing Line
The Finishing Line
 
Conversion Optimization Willa Fogarty
Conversion Optimization Willa FogartyConversion Optimization Willa Fogarty
Conversion Optimization Willa Fogarty
 
Case Study by Jon Warden and Shruti Ahuja of Haymarket - Optimizely Experienc...
Case Study by Jon Warden and Shruti Ahuja of Haymarket - Optimizely Experienc...Case Study by Jon Warden and Shruti Ahuja of Haymarket - Optimizely Experienc...
Case Study by Jon Warden and Shruti Ahuja of Haymarket - Optimizely Experienc...
 
Google Analytics Powerups and Smartcuts
Google Analytics Powerups and Smartcuts Google Analytics Powerups and Smartcuts
Google Analytics Powerups and Smartcuts
 
Patrick McKenzie Opticon 2014: Advanced A/B Testing
Patrick McKenzie Opticon 2014: Advanced A/B TestingPatrick McKenzie Opticon 2014: Advanced A/B Testing
Patrick McKenzie Opticon 2014: Advanced A/B Testing
 
Customer Development Fast Protyping
Customer Development Fast ProtypingCustomer Development Fast Protyping
Customer Development Fast Protyping
 
Growing Up The Right Way: An Example of How to Scale When You're Aren't a Sta...
Growing Up The Right Way: An Example of How to Scale When You're Aren't a Sta...Growing Up The Right Way: An Example of How to Scale When You're Aren't a Sta...
Growing Up The Right Way: An Example of How to Scale When You're Aren't a Sta...
 
Craig Sullivan - Keynote speaker summary & final thoughts - Conversion Hotel ...
Craig Sullivan - Keynote speaker summary & final thoughts - Conversion Hotel ...Craig Sullivan - Keynote speaker summary & final thoughts - Conversion Hotel ...
Craig Sullivan - Keynote speaker summary & final thoughts - Conversion Hotel ...
 
Lean Startup 301
Lean Startup 301Lean Startup 301
Lean Startup 301
 
iSG Webinar – AB Testing: The most important thing you’re NOT doing
iSG Webinar – AB Testing: The most important thing you’re NOT doingiSG Webinar – AB Testing: The most important thing you’re NOT doing
iSG Webinar – AB Testing: The most important thing you’re NOT doing
 
Research and Discovery Tools for Experimentation - 17 Apr 2024 - v 2.3 (1).pdf
Research and Discovery Tools for Experimentation - 17 Apr 2024 - v 2.3 (1).pdfResearch and Discovery Tools for Experimentation - 17 Apr 2024 - v 2.3 (1).pdf
Research and Discovery Tools for Experimentation - 17 Apr 2024 - v 2.3 (1).pdf
 
Management Consulting Productivity Hacks
Management Consulting Productivity HacksManagement Consulting Productivity Hacks
Management Consulting Productivity Hacks
 
OEE Interphex New York 2013
OEE Interphex New York 2013OEE Interphex New York 2013
OEE Interphex New York 2013
 

Más de Craig Sullivan

Why does my Mobile Conversion rate suck? 19 Sep 2013 @ Conversion Thursday #...
Why does my Mobile Conversion rate suck?  19 Sep 2013 @ Conversion Thursday #...Why does my Mobile Conversion rate suck?  19 Sep 2013 @ Conversion Thursday #...
Why does my Mobile Conversion rate suck? 19 Sep 2013 @ Conversion Thursday #...Craig Sullivan
 
Elite Camp 2013 - Estonia
Elite Camp 2013 - EstoniaElite Camp 2013 - Estonia
Elite Camp 2013 - EstoniaCraig Sullivan
 
Conversionista : Conversion manager course - Stockholm 20 march 2013
Conversionista : Conversion manager course  - Stockholm 20 march 2013Conversionista : Conversion manager course  - Stockholm 20 march 2013
Conversionista : Conversion manager course - Stockholm 20 march 2013Craig Sullivan
 
3 Optimisation Decks : WAW Copenhagen - 27 Feb 2013
3 Optimisation Decks : WAW Copenhagen - 27 Feb 20133 Optimisation Decks : WAW Copenhagen - 27 Feb 2013
3 Optimisation Decks : WAW Copenhagen - 27 Feb 2013Craig Sullivan
 
Measure camp tools of the cro rabble
Measure camp   tools of the cro rabbleMeasure camp   tools of the cro rabble
Measure camp tools of the cro rabbleCraig Sullivan
 
5 cro tools that i can't live without
5 cro tools that i can't live without5 cro tools that i can't live without
5 cro tools that i can't live withoutCraig Sullivan
 
eMetrics Stockholm - What the F*** is wrong with my conversion?
eMetrics Stockholm - What the F*** is wrong with my conversion?eMetrics Stockholm - What the F*** is wrong with my conversion?
eMetrics Stockholm - What the F*** is wrong with my conversion?Craig Sullivan
 

Más de Craig Sullivan (7)

Why does my Mobile Conversion rate suck? 19 Sep 2013 @ Conversion Thursday #...
Why does my Mobile Conversion rate suck?  19 Sep 2013 @ Conversion Thursday #...Why does my Mobile Conversion rate suck?  19 Sep 2013 @ Conversion Thursday #...
Why does my Mobile Conversion rate suck? 19 Sep 2013 @ Conversion Thursday #...
 
Elite Camp 2013 - Estonia
Elite Camp 2013 - EstoniaElite Camp 2013 - Estonia
Elite Camp 2013 - Estonia
 
Conversionista : Conversion manager course - Stockholm 20 march 2013
Conversionista : Conversion manager course  - Stockholm 20 march 2013Conversionista : Conversion manager course  - Stockholm 20 march 2013
Conversionista : Conversion manager course - Stockholm 20 march 2013
 
3 Optimisation Decks : WAW Copenhagen - 27 Feb 2013
3 Optimisation Decks : WAW Copenhagen - 27 Feb 20133 Optimisation Decks : WAW Copenhagen - 27 Feb 2013
3 Optimisation Decks : WAW Copenhagen - 27 Feb 2013
 
Measure camp tools of the cro rabble
Measure camp   tools of the cro rabbleMeasure camp   tools of the cro rabble
Measure camp tools of the cro rabble
 
5 cro tools that i can't live without
5 cro tools that i can't live without5 cro tools that i can't live without
5 cro tools that i can't live without
 
eMetrics Stockholm - What the F*** is wrong with my conversion?
eMetrics Stockholm - What the F*** is wrong with my conversion?eMetrics Stockholm - What the F*** is wrong with my conversion?
eMetrics Stockholm - What the F*** is wrong with my conversion?
 

#Measurecamp : 18 Simple Ways to F*** up Your AB Testing

  • 1. 18 simple ways to fuck up your AB testing 28th March 2014 @OptimiseOrDie
  • 2. @OptimiseOrDie • UX and Analytics (1999) • User Centred Design (2001) • Agile, Startups, No budget (2003) • Funnel optimisation (2004) • Multivariate & A/B (2005) • Conversion Optimisation (2005) • Persuasive Copywriting (2006) • Joined Twitter (2007) • Lean UX (2008) • Holistic Optimisation (2009) Was : Group eBusiness Manager, Belron Now : Spareroom.co.uk
  • 3. @OptimiseOrDie Timeline Tested stupid ideas, lots Most AB or MVT tests are bullshit Discovered AB testing Triage, Triangulation, Prioritisation, Maths Zen Plumbing AB Test Hype Cycle
  • 4. Craig’s Cynical Quadrant Improves revenue Improves UX YesNo No Yes Client delighted (and fires you for another UX agency) Client fucking delighted Client absolutely fucking furious Client fires you (then wins an award for your work)
  • 5. #1 : You’re doing it in the wrong place @OptimiseOrDie
  • 6. #1 : You’re doing it in the wrong place There are 4 areas a CRO expert always looks at: 1. Inbound attrition (medium, source, landing page, keyword, intent and many more…) 2. Key conversion points (product, basket, registration) 3. Processes and steps (forms, logins, registration, checkout) 4. Layers of engagement (search, category, product, add) 1. Use visitor flow reports for attrition – very useful. 2. For key conversion points, look at loss rates & interactions 3. Processes and steps – look at funnels or make your own 4. Layers and engagement – make a ring model @OptimiseOrDie
  • 8. Examples – 16-25Railcard.co.uk Bounce Login to Account Content Engage Start Application Type and Details Eligibility Photo Complete @OptimiseOrDie
  • 9. Examples – Guide Dogs Bounce Content Engage Donation Pathway Donation Page Starts process Funnel steps Complete @OptimiseOrDie
  • 10. Within a layer Page 1 Page 2 Page 3 Page 4 Page 5 Exit Deeper Layer Email LikeContact Wishlist Micro Conversions @OptimiseOrDie
  • 11. #1 : You’re doing it in the wrong place • Get to know the flow and loss (leaks) inbound, inside and through key processes or conversion points. • Once you know the key steps you’re losing people at and how much traffic you have – make a money model. • Let’s say 1,000 people see the page a month. Of those, 20% (200) convert to checkout. • Estimate the influence your test can bring. How much money or KPI improvement would a 10% lift in the checkouts deliver? • Congratulations – you’ve now built the worlds first IT plan with a return on investment estimate attached! • I’ll talk more about prioritising later – but a good real world analogy for you to use: @OptimiseOrDie
  • 12. Think like a store owner! If you can’t refurbish the entire store, which floors or departments will you invest in optimising? Wherever there is: • Footfall • Low return • Opportunity @OptimiseOrDie
  • 13. Insight - Inputs #FAIL Competitor copying Guessing Dice rolling An article the CEO read Competitor change Panic Ego Opinion Cherished notions Marketing whims Cosmic rays Not ‘on brand’ enough IT inflexibility Internal company needs Some dumbass consultant Shiny feature blindness Knee jerk reactons #2 : Your hypothesis is crap! @OptimiseOrDie
  • 14. Insight - Inputs Insight Segmentation Surveys Sales and Call Centre Session Replay Social analytics Customer contact Eye tracking Usability testing Forms analytics Search analytics Voice of Customer Market research A/B and MVT testing Big & unstructured data Web analytics Competitor evalsCustomer services #2 : These are the inputs you need… @OptimiseOrDie
  • 15. #2 : Solutions • You need multiple tool inputs – Tool decks are here : www.slideshare.net/sullivac • Usability testing and User facing teams – If you’re not doing these properly, you’re hosed • Session replay tools provide vital input – Get vital additional customer evidence • Simple page Analytics don’t cut it – Invest in your analytics, especially event tracking • Ego, Opinion, Cherished notions – fill gaps – Fill these vacuums with insights and data • Champion the user – Give them a chair at every meeting @OptimiseOrDie
  • 16. We believe that doing [A] for People [B] will make outcome [C] happen. We’ll know this when we observe data [D] and obtain feedback [E]. (reverse) @OptimiseOrDie
  • 17. #3 : No analytics integration • Investigating problems with tests • Segmentation of results • Tests that fail, flip or move around • Tests that don’t make sense • Broken test setups • What drives the averages you see? @OptimiseOrDie
  • 18. 18 A B B A
  • 19. These Danish porn sites are so hardcore! We’re still waiting for our AB tests to finish! • Use a test length calculator like this one: • visualwebsiteoptimizer.com/ab-split-test-duration/#4 : The test will finish after you die
  • 20. #5 : You don’t test for long enough • The minimum length – 2 business cycles (cross check) – Usually a week, 2 weeks, Month – Always test ‘whole’ not partial cycles – Be aware of multiple cycles – Don’t self stop! – PURCHASE CYCLES – KNOW THEM • How long after that – I aim for a minimum 250 outcomes, ideally 350+ for each ‘creative’ – If you test 4 recipes, that’s 1400 outcomes needed – You should have worked out how long each batch of 350 needs before you start! – 95% confidence or higher is my aim BUT BIG SECRET -> (p values are unreliable) – If you segment, you’ll need more data – It may need a bigger sample if the response rates are similar* – Use a test length calculator but be aware of BARE MINIMUM TO EXPECT – Important insider tip – watch the error bars! The +/- stuff – let’s explain * Stats geeks know I’m glossing over something here. That test time depends on how the two experiments separate in terms of relative performance as well as how volatile the test response is. I’ll talk about this when I record this one! This is why testing similar stuff sux. 20
  • 21. #5 : The tennis court – Let’s say we want to estimate, on average, what height Roger Federer and Nadal hit the ball over the net at. So, let’s start the match: @OptimiseOrDie
  • 22. First Set Federer 6-4 – We start to collect values 62cm +/- 2cm 63.5cm +/- 2cm @OptimiseOrDie
  • 23. Second Set – Nadal 7-6 – Nadal starts sending them low over the net 62cm +/- 1cm 62.5cm +/- 1cm @OptimiseOrDie
  • 24. Final Set Nadal 7-6 – We start to collect values 61.8cm +/- .3cm 62cm +/- .3cm
  • 25. Let’s look at this a different way 62.5cm +/- 1cm @OptimiseOrDie 9.1% ± 0.3 9.3% ± 0.3
  • 26. 62.5cm +/- 1cm @OptimiseOrDie 9.1% ± 0.5 9.3% ± 0.5 9.1% ± 0.2 9.3% ± 0.2 9.1% ± 0.1 9.3% ± 0.1
  • 27. Graph is a range, not a line: 9.1 ± 0.3%9.1 ± 0.9%9.1 ± 1.9%
  • 28. #5 : Summary • The minimum length: – 2 business cycles and > purchase cycle as a minimum, regardless of outcomes. Test for less and you’re cutting. – 250+, prefer 350+ outcomes in each – Error bar separation between creatives – 95%+ confidence (unreliable) • Pay attention to: – Time it will take for the number of ‘recipes’ in the test – The actual footfall to the test – not sitewide numbers – Test results that don’t separate – makes the test longer – This is why you need brave tests – to drive difference – The error bars – the numbers in your AB testing tool are not precise – they’re fuzzy regions that depend on response and sample size. – Sudden changes in test performance or response – Monitor early tests like a chef! @OptimiseOrDie
  • 29. #6 : You peek and jump to conclusions! • Ignore the graphs. Don’t draw conclusions. Don’t dance. Calm down. • Get a feel for the test but don’t do anything yet! • Remember – in A/B - 50% of returning visitors will see a new shiny website! • Until your test has had at least 1 business cycle and 250-350 outcomes, don’t bother even getting excited! • Watching regularly is good though. You’re looking for anything that looks really odd – your analytics person should be checking all the figures until you’re satisfied • All tests move around or show big swings early in the testing cycle. Here is a very high traffic site – it still takes 10 days to start settling. Lower traffic sites will stretch this period further. 29
  • 30. #7 : No QA testing for the AB test?
  • 31. #7 - QA Test or Die! • Over 40% of tests have had QA issues. • It’s very easy to break or bias the testing Browser testing www.crossbrowsertesting.com www.browserstack.com www.spoon.net www.cloudtesting.com www.multibrowserviewer.com www.saucelabs.com Mobile devices www.perfectomobile.com www.deviceanywhere.com www.mobilexweb.com/emulators www.opendevicelab.com @OptimiseOrDie
  • 32. #7 : What other QA testing should I do? • Cross Browser Testing • Testing from several locations (office, home, elsewhere) • Testing the IP filtering is set up • Test tags are firing correctly (analytics and the test tool) • Test as a repeat visitor and check session timeouts • Cross check figures from 2+ sources • Monitor closely from launch, recheck, watch • WATCH FOR BIAS! @OptimiseOrDie
  • 33. #8 : Opportunities are not prioritised Once you have a list of potential test areas, rank them by opportunity vs. effort. The common ranking metrics that I use include: •Opportunity (revenue, impact) •Dev resource •Time to market •Risk / Complexity Make yourself a quadrant diagram and plot them
  • 34. #9 : Your cycles are too slow 0 6 12 18 Months Conversion @OptimiseOrDie
  • 35. #9 : Solutions • Give Priority Boarding for opportunities – The best seats reserved for metric shifters • Release more often to close the gap – More testing resource helps, analytics ‘hawk eye’ • Kaizen – continuous improvement – Others call it JFDI (just f***ing do it) • Make changes AS WELL as tests, basically! – These small things add up • RUSH Hair booking – Over 100 changes – No functional changes at all – 37% improvement • Inbetween product lifecycles? – The added lift for 10 days work, worth 360k @OptimiseOrDie
  • 36. #9 : Make your own cycles @OptimiseOrDie
  • 37. #10 : How do I know when it’s ready? • The hallmarks of a cooked test are: – It’s done at least 1 or preferably 2+ business and at least one if not two purchase cycles – You have at least 250-350 outcomes for each recipe – It’s not moving around hugely at creative or segment level performance – The test results are clear – even if the precise values are not – The intervals are not overlapping (much) – If a test is still moving around, you need to investigate – Always declare on a business cycle boundary – not the middle of a period (this introduces bias) – Don’t declare in the middle of a limited time period advertising campaign (e.g. TV, print, online) – Always test before and after large marketing campaigns (one week on, one week off) 37
  • 38. 38 #11 : Your test fails @OptimiseOrDie
  • 39. #11: Your test fails • Learn from the failure! If you can’t learn from the failure, you’ve designed a crap test. • Next time you design, imagine all your stuff failing. What would you do? If you don’t know or you’re not sure, get it changed so that a negative becomes insightful. • So : failure itself at a creative or variable level should tell you something. • On a failed test, always analyse the segmentation and analytics • One or more segments will be over and under • Check for varied performance • Now add the failure info to your Knowledge Base: • Look at it carefully – what does the failure tell you? Which element do you think drove the failure? • If you know what failed (e.g. making the price bigger) then you have very useful information • You turned the handle the wrong way • Now brainstorm a new test @OptimiseOrDie
  • 40. #12 : The test is ‘about the same’ • Analyse the segmentation • Check the analytics and instrumentation • One or more segments may be over and under • They may be cancelling out – the average is a lie • The segment level performance will help you (beware of small sample sizes) • If you genuinely have a test which failed to move any segments, it’s a crap test – be bolder • This usually happens when it isn’t bold or brave enough in shifting away from the original design, particularly on lower traffic sites • Get testing again! @OptimiseOrDie
  • 41. • There are three reasons it is moving around – Your sample size (outcomes) is still too small – The external traffic mix, customers or reaction has suddenly changed or – Your inbound marketing driven traffic mix is completely volatile (very rare) • Check the sample size • Check all your marketing activity • Check the instrumentation • If no reason, check segmentation #13 : The test keeps moving around @OptimiseOrDie
  • 42. • Something like this can happen: • Check your sample size. If it’s still small, then expect this until the test settles. • If the test does genuinely flip – and quite severely – then something has changed with the traffic mix, the customer base or your advertising. Maybe the PPC budget ran out? Seriously! • To analyse a flipped test, you’ll need to check your segmented data. This is why you have a split testing package AND an analytics system. • The segmented data will help you to identify the source of the shift in response to your test. I rarely get a flipped one and it’s always something changing on me, without being told. The heartless bastards. #14 : The test has flipped on me
  • 43. • No – and this is why: – It’s a waste of time – It’s easier to test and monitor instead – You are eating into test time – Also applies to A/A/B/B testing – A/B/A running at 25%/50%/25% is the best • Read my post here : http://bit.ly/WcI9EZ 43 #15 : Should I run an A/A test first
  • 44. #16 : Nobody feels the test • You promised a 25% rise in checkouts - you only see 2% • Traffic, Advertising, Marketing may have changed • Check they’re using the same precise metrics • Run a calibration exercise • I often leave a 5 or 10% stub running in a test • This tracks old creative once new one goes live • If conversion is also down for that one, BINGO! • Remember – the AB test is an estimate – it doesn’t precisely record future performance • This is why infrequent testing is bad • Always be trying a new test instead of basking in the glory of one you ran 6 months ago. You’re only as good as your next test. @OptimiseOrDie
  • 45. #17 : You forgot about Mobile & Tablet • If you’re AB testing a responsive site, pay attention • Content will break differently on many screens • Know thy users and their devices • Use bango or google analytics to define a test list • Make sure you test mobile devices & viewports • What looks good on your desk may not be for the user • Harder to design cross device tests • You’ll need to segment mobile, tablet & desktop response in the analytics or AB testing package • Your personal phone is not a device mix • Ask me about making your device list • Buy core devices, rent the rest from deviceanywhere.com @OptimiseOrDie
  • 46. • Forget MVT or A/B/N tests – run your numbers • Test things with high impact – don’t be a wuss! • Use UX, Session Replay to aid insight • Run a task gap survey (4Q style) • Run a dropped basket survey (LF style) • Run a general survey + check social + other sites • Run sitewide tests that appear on all pages or large clusters of pages – • UVPs (“We are a cool brand”), USPs (“Free returns!”), UCPs (“10% off today”). • Headers, Footers, Nudge Bars, USP bars, footer changes, Navigation, Product pages, Delivery info etc. #18 : Oh shit – no traffic!
  • 47. • If small volumes, contact customers – reach out. • If data volumes aren’t there, there are still customers! • Drive design from levers you can apply – game the system • Pick clean and simple clusters of change (hypothesis driven) • Use a goal at an earlier ring stage or funnel step • Beware of using clickthroughs when attrition is high on the other side • Try before and after testing on identical time periods (measure in analytics model) • Be careful about small sample sizes (<100 outcomes) • Are you working automated emails? • Fix JFDI, performance and UX issues too! #18 : Oh shit – no traffic
  • 48. Top F***ups for 2014 1. Testing in the wrong place 2. Your hypothesis inputs are crap 3. No analytics integration 4. Your test will finish after you die 5. You don’t test for long enough 6. You peek before it’s ready 7. No QA for your split test 8. Opportunities are not prioritised 9. Testing cycles are too slow 10. You don’t know when tests are ready 11. Your test fails 12. The test is ‘about the same’ 13. Test flips behaviour 14. Test keeps moving around 15. You run an A/A test and waste time 16. Nobody ‘feels’ the test 17. You forgot you were responsive 18. You forgot you had no traffic @OptimiseOrDie
  • 49. Is there a way to fix this then? 49 Conversion Heroes! @OptimiseOrDie
  • 50. Email Twitter : sullivac@gmail.com : @OptimiseOrDie : linkd.in/pvrg14 Slides uploaded to SLIDESHARE.NETSULLIVAC 50