Download - Testing for Success

Transcript
Page 1: Testing for Success

>  Tes&ng  for  Success  <  Elements  of  a  Successful  Tes0ng  

Program  

Page 2: Testing for Success

>  Agenda  

§ Why  Test?              § Problem  Diagnosis  § Deciding  what  to  Test      § Test  Execu0on  and  Measurement  § Test  Repor0ng  

December  2011   ©  Datalicious  Pty  Ltd   2  

Page 3: Testing for Success

>  Why  Test?  

101011010010010010101111010010010101010100001011111001010101010100101011001100010100101001101101001101001010100111001010010010101001001010010100100101001111101010100101001001001010  

December  2011   ©  Datalicious  Pty  Ltd   3  

Page 4: Testing for Success

1. Why  does  your  business/organisa0on  exist?  

2.  How  can  your  business/organisa0on  improve?  

December  2011   ©  Datalicious  Pty  Ltd   4  

EVERYONE’S  GOT  AN  OPINION  

Page 5: Testing for Success

>  Why  Test?  

1.  Systema0c  Innova0on  2.  Avoid  costly  mistakes  3.  Know  why  things  go  right,  know  why  things  

go  wrong  4.  BeRer  employee  engagement  

§  Requires  planning  and  governance!  

December  2011   ©  Datalicious  Pty  Ltd   5  

Page 6: Testing for Success

>  Problem  Diagnosis  

101011010010010010101111010010010101010100001011111001010101010100101011001100010100101001101101001101001010100111001010010010101001001010010100100101001111101010100101001001001010  

December  2011   ©  Datalicious  Pty  Ltd   6  

Page 7: Testing for Success

>  What  is  the  business  problem?  

December  2011   ©  Datalicious  Pty  Ltd   7  

Analy&cs  and  metrics  frameworks  

Acquisi0on   Up-­‐Sell   Reten0on   Advocacy  

Page 8: Testing for Success

>  Case  Study  

December  2011   ©  Datalicious  Pty  Ltd   8  

Page 9: Testing for Success

>  Further  Diagnosis  

December  2011   9  

PROBLEM:  Sales  through  online  

Not  enough  site  traffic  

©  Datalicious  Pty  Ltd  

High  home  page  bounce  rate  

Low  conversion  on  product  page  

Checkout  fallout  

Page 10: Testing for Success

>  Further  Diagnosis  II  

©  Datalicious  Pty  Ltd   10  December  2011  

Source:  www.feng-­‐gui.com  

Page 11: Testing for Success

>  Some&mes  the  small  things  count  

December  2011   ©  Datalicious  Pty  Ltd   11  

Page 12: Testing for Success

December  2011   ©  Datalicious  Pty  Ltd   12  

>  Further  diagnosis  III  

Wrong  message?  Wrong  channel?  Wrong  person?  Wrong  0me?  

Page 13: Testing for Success

>  Tes&ng  as  risk  mi&ga&on  

December  2011   ©  Datalicious  Pty  Ltd   13  

Roll-­‐out  Channel    

Press   TV   Radio   Outdoor  

Test  Channel  

eDM/DM  Offer,  

Crea&ve,  Call-­‐to-­‐Ac&on  

Call-­‐to-­‐Ac&on  

Offer,  Call-­‐to-­‐Ac&on  

Offer,  Call-­‐to-­‐Ac&on  

Paid  Search  

Offer   Offer   Offer   Offer  

Display  Media  

-­‐   Crea&ve   Offer,  Call-­‐to  Ac&on  

Crea&ve,  Offer,  Call-­‐to  Ac&on  

Page 14: Testing for Success

>  Tes&ng  as  standard  prac&ce  

December  2011   ©  Datalicious  Pty  Ltd   14  

 %  Uplib  in  Sales  

Test  Market  Control  Market  (no  ATL)  

 Time  

Page 15: Testing for Success

>  Deciding  what  to  Test  

101011010010010010101111010010010101010100001011111001010101010100101011001100010100101001101101001101001010100111001010010010101001001010010100100101001111101010100101001001001010  

December  2011   ©  Datalicious  Pty  Ltd   15  

Page 16: Testing for Success

December  2011   ©  Datalicious  Pty  Ltd   16  

Don’t  reinvent  the  wheel  

Page 17: Testing for Success

©  Datalicious  Pty  Ltd  December  2011   17  

>  What  are  the  solu&on(s)?  

©  Datalicious  Pty  Ltd   17  

Page 18: Testing for Success

December  2011   ©  Datalicious  Pty  Ltd   18  

What  are  your  visitors  trying  to  achieve  by  visi2ng  your  site?  

>  Consumer  Empathy  

Page 19: Testing for Success

>  Consumer  Empathy  

December  2011   ©  Datalicious  Pty  Ltd   19  

1.   Make  it  visible  –  People  can’t  convert  if  they  can’t  find  your  

‘Buy  Now’  buRon  

2.   Make  it  relevant  –  Need  to  resolve  consumer  reserva0ons/

ques0ons  

3.   Make  it  easy  –  Easy  naviga0on,  easy  form  comple0on,  easy  to  

read,  quick  page  load  

Page 20: Testing for Success

>  Start  with  the  basics…  1.  The  headline  – Have  a  headline!  – Headline  should  be  concrete  – Headline  should  be  first  thing  visitors  look  at  

2.  Call  to  ac&on  – Don’t  have  too  many  calls  to  ac0on  – Have  an  ac0onable  call  to  ac0on  – Have  a  big,  prominent,  visible  call  to  ac0on  

3.  Social  proof  –  Logos,  number  of  users,  tes0monials,    case  studies,  media  coverage,  etc  

December  2011   ©  Datalicious  Pty  Ltd   20  

Page 21: Testing for Success

December  2011   ©  Datalicious  Pty  Ltd   21  

>  Start  with  the  basics…  

Page 22: Testing for Success

>  Case  Study  

December  2011   ©  Datalicious  Pty  Ltd   22  

Page 23: Testing for Success

>  Further  Examples  

December  2011   ©  Datalicious  Pty  Ltd   23  

TEST  A   EXISTING  

Page 24: Testing for Success

>  Further  Examples  

December  2011   ©  Datalicious  Pty  Ltd   24  

EXISTING  

TEST  

Page 25: Testing for Success

>  Deciding  What  to  Test  

Test  Selec0on  Checklist  

§  Is  the  measurement  infrastructure  in  place  already?  §  Can  I  readily  execute  the  solu0on?  §  Do  I  have  enough  sample  to  draw  valid  conclusions?  §  Will  this  prove  the  value  of  tes0ng  in  the  business?  

December  2011   ©  Datalicious  Pty  Ltd   25  

[          ]            ✔  [          ]            ✔  [          ]            ✔  [          ]            ✔  

Page 26: Testing for Success

>    Do  you  have  the  repor&ng?  

December  2011   ©  Datalicious  Pty  Ltd   26  

Test  Channel    

ATL   DM   eDM   Online  

Response  Channel  

Online  

Mailroom  

Call  Centre  

Bricks  &  Mortar  

Channels  in  Aggregate   ✔  

✔  

✔  ✔  

For  each  of  Segment  X,  Y  and  Z...  

Page 27: Testing for Success

>  Offline  conversions  from  online  

December  2011   ©  Datalicious  Pty  Ltd   27  

Cookie  

Website.com  Research  

Phone  Orders  

Retail  Orders  

Online  Orders  

Website.com  Research  

Website.com  Research  

Online  Order  Confirma&on  

Virtual  Order  Confirma&on  

Virtual  Order  Confirma&on  

Virtual  Order  Confirma&on  

@  

@  

@  

Cookie  Cookie  

Online  Ad  Campaign  

Tying  offline  conversions  back  to  online  campaign  and  research  behavior  using  standard  cookie  technology  by  triggering  virtual  online  order  confirma0on  pages  for  offline  sales  using  email  receipts.  

Page 28: Testing for Success

>  Search  call  to  ac&on  for  offline    

December  2011   ©  Datalicious  Pty  Ltd   28  

Page 29: Testing for Success

>  OTP  Response  

December  2011   ©  Datalicious  Pty  Ltd   29  

– Different  numbers  for  different  media  channels  – Different  numbers  for  different  product  categories  

– Different  numbers  for  different  conversion  steps  – Call  origin  becoming  useful  to  shape  call  script  – Feasible  to  pause  numbers  to  improve  integrity  

Page 30: Testing for Success

>  Whose  help  do  you  need?  

December  2011   ©  Datalicious  Pty  Ltd   30  

Technology/IT  

Analytics!

Creative Agency

UX Agency Your boss, Your boss’ boss

Customer Contact Management

Page 31: Testing for Success

>  Proving  the  Value  

December  2011   ©  Datalicious  Pty  Ltd   31  

GO  BIG  

Page 32: Testing for Success

>  How  much  sample  do  I  need?  

December  2011   ©  Datalicious  Pty  Ltd   32  

#  on  Segments,  #  of  Treatments  

BAU/Baseline  Conversion  Rate  

Time  in  Market  [Digital  Only]  

Expected  Δ  in  Conversion  n  

Page 33: Testing for Success

>  Sta&s&cal  Significance  

December  2011   ©  Datalicious  Pty  Ltd   33  

Q.  How  much  am  I  willing  to  accept  that  the        difference  in  the  results  between  my  test  group  and  control  group  may  have  been  due  to  chance?  

 A.   Not  much.  I  want  to  be  confident  that  if  I  

repeated  the  test  100  &mes,  then  I  would  observe  this  difference  95  &mes.    

 This  is  ‘95%  confidence’  

Page 34: Testing for Success

>  Type  I  and  Type  II  Error  

December  2011   ©  Datalicious  Pty  Ltd   34  

Type  I:    Accept  result  to  be  true  when  it’s    actually  false  (false  posi&ves)  

 Type  II:  Accept  result  to  be  false  when  it’s    

 actually  true  (false  nega&ves)  

Page 35: Testing for Success

>  Es&ma&ng  Sample  Size  (%s)  

December  2011   ©  Datalicious  Pty  Ltd   35  

n = 2(1.645+1.282) *

p1(1− p1)+ p2 (1− p2 )Δ2

#

$%

&

'(

Where:    n    =    es0mated  sample  size  for  each  group    p1  =    expected  conversion  rate  for  your  test  treatment    p2  =    expected  conversion  rate  for  your  control  treatment    Δ    =    expected  minimum  percentage  point  difference  between  test        and  control  results      

 The  value  of  1.645  reflects  that  we  accept  Type  I  error  probability  of  .05    The  value  of  1.282  reflects  that  we  accept  Type  II  error  probability  of  .10    

Page 36: Testing for Success

>  Es&ma&ng  Sample  Size  (%s)  

December  2011   ©  Datalicious  Pty  Ltd   36  

n = 2(1.645+1.282) *

0.025*0.975+ 0.030*0.9700.0052

!

"#

$

%&

Typical  Champion  (control)  vs.  Challenger  (test)  A|B  test,  typical  champion  response  rate  of  2.5%.    

•  Only  going  to  replace  Champion  with  Challenger  if  Challenger  response  rate  is  3.0%  (0.5%  is  a  meaningful  difference)    

 

Sample  size  =  18,326  For  each  of  the  Champion  and  Challenger  groups    If  1.0%  our  meaningful  difference  then  sample  size  is  only  5,378  

Page 37: Testing for Success

>  Es&ma&ng  Sample  Size  ($s)  

December  2011   ©  Datalicious  Pty  Ltd   37  

n = (1.645+1.282)2 *(s1

2 + s22 )

Δ2

Where:    n    =    number  of  observa0ons  for  each  group    s1  =    expected  standard  devia0on  of  value  for  your  test  treatment    s2  =    expected  standard  devia0on  of  value  for  your  control  treatment    Δ    =    expected  minimum  difference  in  value  between  test        and  control  results      

 The  value  of  1.645  reflects  an  accepted  Type  I  error  probability  of  .05    The  value  of  1.282  reflects  an  accepted  Type  II  error  probability  of  .10    

Page 38: Testing for Success

>  Standard  Devia&on  

December  2011   ©  Datalicious  Pty  Ltd   38  

Where:    n    =    number  of  observa0ons    xi  =    the  result  for  the  ith  observa0on    x  =    mean  (average)  for  your  data  

Standard  devia0on  is  measure  of  the  variability  of  your  results,  whether  some  your  results  are  quite  different  to  your  mean  (average)  result  or  whether  they  are  quite  similar.  

s =(xi − x )

i=1

n

∑n−1

Page 39: Testing for Success

>  Es&ma&ng  Sample  Size  ($s)  

December  2011   ©  Datalicious  Pty  Ltd   39  

Typical  Champion  (control)  vs.  Challenger  (test)  A|B  test,  typical  champion  mean  response  value  of  $20,  typical  response  rate  of  5%    

•  Only  going  to  replace  Champion  with  Challenger  if  Challenger  mean  response  value  is  is  $30  ($10  is  a  meaningful  difference)  

•  Standard  devia0on  of  Champion  results  is  $5  (based  on  past  results).  We’ll  assume  the  same  for  the  Challenger.    

     

n = (1.645+1.282)2 *(52 + 52 )

102Number  of  observa0ons  =  4.3  (~5)  for  each  of  the  Champion  and  Challenger  groups.    Then  divide  through  with  the  expected  response  rate  to  get  minimum  sample  size  of  86  for  each  of  Challenger  and  Control  groups  (4.3/0.05)  

Page 40: Testing for Success

>  Further  Complexity  I  

December  2011   ©  Datalicious  Pty  Ltd   40  

If  we  wanted  to  test  the  performance  of  Challenger  vs.  Champion  for  different  segments  of  consumers:  

Using  same  assump0ons  as  in  earlier  example  need  18,326  per  cell,  18,326*6=109,956  in  total  .    

Response  Rate  

Champion   Challenger  

Segment  

A   %   %  

B   %   %  

C   %   %  

Page 41: Testing for Success

>  Further  Complexity  II  

December  2011   ©  Datalicious  Pty  Ltd   41  

If  we  wanted  to  test  the  performance  of  Challenger  vs.  Champion  for  difference  segments  of  consumers  AND  had  3  different  types  of  Champion  crea0ve:  

Using  same  assump0ons  as  in  earlier  example  need  18,326  per  cell,  18,326*12=219,912  in  total.    

Response  Rate  

Champion   Challenger  #1  

Challenger  #2  

Challenger  #3  

Segment  

A   %   %   %   %  

B   %   %   %   %  

C   %   %   %   %  

Page 42: Testing for Success

>  Further  Complexity  III  

December  2011   ©  Datalicious  Pty  Ltd   42  

If  we  wanted  to  test  the  performance  of  Challenger  crea0ve  that  was  specifically  customised  for  difference  segments  of  consumers,  then  we’re  actually  only  running  6  tests  

Using  same  assump0ons  as  in  earlier  example  need  18,326  per  cell,  18,326*6=109.956  in  total.    

Response  Rate  

Champion   Challenger  #1  

Challenger  #2  

Challenger  #3  

Segment  

A   %   %  

B   %   %  

C   %   %  

Page 43: Testing for Success

>  Mul&variate  Tes&ng  

December  2011   ©  Datalicious  Pty  Ltd   43  

Mul0variate  Tes0ng  (commonly  called  MVT)  is  a  term  used  for  tes0ng  different  varia0ons  of  typical  elements  of  a  landing  page,  direct  mail  leRer,  etc.    The  aim  is  to  determine  which  combina0on  delivers  the  best  result.  

Element  #1:  Prominent  headline  

Element  #2:    Call  to  ac0on  

Suppor0ng    content  

Element  #3:  Social  proof  /  trust  

Terms  and  condi0ons  

§  Element  #1  –  2  varia0ons  (1  exis0ng,  1  new)  

§  Element  #2  –  2  varia0ons  (1  exis0ng,  1  new)  

§  Element  #3:  –  2  varia0ons  (1  exis0ng,  1  new)  

Page 44: Testing for Success

>  MVT  –  Full  Factorial  

December  2011   ©  Datalicious  Pty  Ltd   44  

A  full  factorial  design  requires  every  unique  combina0on  of  page  elements  and  can  therefore  be  very  sample  hungry.    

Element  

Headline   Call  to  Ac&on   Social  Proof  

Treatment  

1   H1   CTA1   SP1  

2   H1   CTA1   SP2  

3   H1   CTA2   SP1  

4   H1   CTA2   SP2  

5   H2   CTA1   SP1  

6   H2   CTA1   SP2  

7   H2   CTA2   SP1  

8   H2   CTA2   SP2  

To  calculate  the  number  of  treatments  just  need  to  mul0ply  the  number  of  varia0ons  for  each  factor  together:    2  x  2  x  2  =    8    

Page 45: Testing for Success

>  MVT  –  Frac&onal  Factorial  

December  2011   ©  Datalicious  Pty  Ltd   45  

The  alterna0ve  is  called  a  frac0onal  factorial  design  which  is  some  smaller  set  of  elements  combina0ons.  The  design  should  be  ‘balanced’  -­‐  every  varia0on  is  tested  the  same  number  of  0mes  and  each  combina0on  of  varia0ons  occurs  the  same  number  of  0mes.  

Element  

Headline   Call  to  Ac&on   Social  Proof  

Treatment  

1  

2   H1   CTA1   SP2  

3   H1   CTA2   SP1  

4  

5   H2   CTA1   SP1  

6  

7  

8   H2   CTA2   SP2  

Reduced  sample  requirements  4x18,326=73,304  

Page 46: Testing for Success

>  Layout  Before  Content  §  Phase  #1:  A|B  test  

–  Test  the  same  landing  page  content  in  completely  different  layouts  

§  Phase  #2:  MV  test  –  Then  test  different  content  element  combina0ons  within  the  winning  layout  

§  Phase  #3:  MV  test  (if  req’d)  –  Test  with  reduced  set  of  elements  

December  2011   ©  Datalicious  Pty  Ltd   46  

Element  #1:  Prominent  headline  

Element  #2:    Call  to  ac0on  

Suppor0ng    content  

Element  #3:  Social  proof  /  trust  

Terms  and  condi0ons  

Page 47: Testing for Success

>  Case  Study  

December  2011   ©  Datalicious  Pty  Ltd   47  

§  Yes,  the  measurement  infrastructure  is  in  place  §  I  can  readily  execute  the  test  design  §  I  have  enough  sample  to  draw  valid  conclusions  §  Yes,  this  design  will  prove  the  value  of  tes0ng  in  my  

business  

Page 48: Testing for Success

>  Execu&on  &  Measurement  

101011010010010010101111010010010101010100001011111001010101010100101011001100010100101001101101001101001010100111001010010010101001001010010100100101001111101010100101001001001010  

December  2011   ©  Datalicious  Pty  Ltd   48  

Page 49: Testing for Success

December  2011   ©  Datalicious  Pty  Ltd   49  

Before  you  leap…  

Page 50: Testing for Success

>  Sample  Selec&on  

§  Each  sample  needs  to  be  alike  in  terms  of  their  predisposi0on  to  conversion  

December  2011   ©  Datalicious  Pty  Ltd   50  

TEST   CONTROL  

Conversion:  low  rate  credit  card  applica0on  form  comple0on  

18-­‐34  Mostly  Male  

Mostly  Low  Income  

35-­‐64  Mostly  Female  

Mostly  High  Income  

Page 51: Testing for Success

>  Timing  is  Important  

December  2011   ©  Datalicious  Pty  Ltd   51  

Sales  

 Time  

 ‘Burst’  Non  BAU  ATL  Campaign  

Ideal  Test  Window  

Page 52: Testing for Success

>  A|A  Tes&ng  

December  2011   ©  Datalicious  Pty  Ltd   52  

§  Set  a  test  that  splits  your  visitors  50/50  between  the  same  treatment  – Check  that  sample  sizes  are  actually  50/50  –  Is  there  should  be  no  difference  in  your  conversion  rates  

– Are  volumes  of  conversions  matching  other  repor0ng?  

Page 53: Testing for Success

>  Measuring  your  performance  

December  2011   ©  Datalicious  Pty  Ltd   53  

§  Propor0ons  (conversion  rates)  § Means  (average  $s)  §  Variability  of  Means  (standard  devia0on)  

 §  Use  confidence  intervals  

Would  my  winning  treatment  s2ll  be  the  winner  across  all  my  customers/visitors/consumers?    

Page 54: Testing for Success

>  Confidence  Intervals  

December  2011   ©  Datalicious  Pty  Ltd   54  

Conversio

n  Ra

te  

 Treatments  A   B   C  

Revenu

e  pe

r  Re

spon

se  

 Treatments  A   B   C  

Page 55: Testing for Success

>  Confidence  Intervals  

December  2011   ©  Datalicious  Pty  Ltd   55  

Page 56: Testing for Success

>  Confidence  Interval  (%s)  

December  2011   ©  Datalicious  Pty  Ltd   56  

Where:    p    =    response  rate    n  =    sample  size  for  treatment  

 The  value  of  1.96  reflects  a  95%  confidence  level  

p̂±1.96* p̂(1− p̂)n

^  

Page 57: Testing for Success

>  Confidence  Interval  Es&ma&on  

December  2011   ©  Datalicious  Pty  Ltd   57  

1.7%±1.96* .017(1−.017)60850

Typical  Champion  (control)  vs.  Challenger  (test)  A|B  Test  

Treatment  

Champion   Challenger  

Mailed   60850   52812  

Responses   1055   455  

Response  Rate   1.7   0.9  

0.9%±1.96* .009(1−.009)52812

1.7%± 0.10% 0.9%± 0.08%

1.69%  ≤  Champion  ≤    1.71%   0.82%  ≤  Challenger  ≤    0.98%  

Page 58: Testing for Success

>  Confidence  Interval  Es&ma&on  

December  2011   ©  Datalicious  Pty  Ltd   58  

p1 − p2 ±1.96*p1(1− p1)

n1+p2 (1− p2 )

n2

Where:    p1    =    response  rate  for  challenger    p2    =    response  rate  for  champion      n1  =    sample  size  for  challenger    n2  =    sample  size  for  challenger  

 The  value  of  1.96  reflects  a  95%  confidence  level  

Page 59: Testing for Success

>  Confidence  Interval  Es&ma&on  

December  2011   ©  Datalicious  Pty  Ltd   59  

0.9−1.7±1.96* .009(1−.009)52812

+.017(1−.017)60850

Typical  Champion  (control)  vs.  Challenger  (test)  A|B  Test  

Treatment  

Champion   Challenger  

Mailed   60850   52812  

Responses   1055   455  

Response  Rate   1.7   0.9  

−0.8± 0.13-­‐0.93%  ≤  Difference  Between  Challenger  and  Champion  ≤    -­‐0.67%  

Page 60: Testing for Success

>  Control  Group  Sample  Size  

December  2011   ©  Datalicious  Pty  Ltd   60  

p1 − p2 ±1.96*p1(1− p1)

n1+p2 (1− p2 )

n2

Where:    nc    =    sample  size  for  control  group    nt    =    sample  size  for  test  group    pc  =    forecast  response  rate  for  control  group    nt  =    forecast  response  rate  for  test  group    m  =    desired  level  of  precision  (%  that  is  a  meaningful  difference)    

 The  value  of  1.96  reflects  a  95%  confidence  level  

nc =pc (1− pc )

m1.96"

#$

%

&'2

−pt (1− pt )

nt

Rearranged:  

Page 61: Testing for Success

>  Control  Group  Sample  Size  

December  2011   ©  Datalicious  Pty  Ltd   61  

nc =.01(1−.01)

.011.96"

#$

%

&'2

−.02(.02−.02)40, 000

We  have  50,000  customers  that  we  could  include  in  our  test  design,  what  would  our  control  sample  need  to  be  if  we  tested  40,000  customers,  our  

‘natural’  cross-­‐sell  rate  was  1.0%  and  an  incremental  response  rate  of  1.0%  points  would  be  deemed  to  be  meaningful?  

This  result  suggests  we  could  actually  test  more  of  our  available  customer  base  than  we  might  have  ini0ally  expected  (~40,600).  

nc = 387

Page 62: Testing for Success

>  Confidence  intervals  ($s)  

December  2011   ©  Datalicious  Pty  Ltd   62  

Where:    x    =    mean  revenue  among  treatment  responders    s  =    standard  devia0on  of  revenue  among  some  treatment’s  responders    n  =    number  of  responders  to  the  treatment  

 The  value  of  1.96  reflects  a  95%  level  of  confidence.    

x ±1.96* sn

Page 63: Testing for Success

>  Standard  Devia&on  (reminder)  

December  2011   ©  Datalicious  Pty  Ltd   63  

Where:    n    =    number  of  observa0ons    xi  =    the  result  for  the  ith  observa0on    x  =    mean  (average)  for  your  data  

Standard  devia0on  is  measure  of  the  variability  of  your  results,  whether  some  your  results  are  quite  different  to  your  mean  (average)  result  or  whether  they  are  quite  similar.  

s =(xi − x )

i=1

n

∑n−1

Page 64: Testing for Success

>  Confidence  intervals  ($s)  

December  2011   ©  Datalicious  Pty  Ltd   64  

Where:    x1  =    mean  value  among  among  responders  to  a  treatment    x2  =    mean  value  among  among  responders  to  a  different  treatment      s1  =    std.  dev.  of  value  among  one  treatment’s  responders    s2  =    std.  dev.  of  value  among  the  other  treatment’s  responders  n1  =    number  of  responders  to  the  treatment    n2  =    number  of  responders  to  the  other  treatment  

 The  value  of  1.96  reflects  a  95%  level  of  confidence.  

n1  and  n2  is  sufficiently  large  to  es0mate  the  std.  dev.  in  the  popula0on  with  the  std.  dev.  of  the  sample.  

x1 − x2 ±1.96*s12

n1+s22

n2

Page 65: Testing for Success

>  Confidence  intervals  ($s)  

December  2011   ©  Datalicious  Pty  Ltd   65  

Typical  Champion  (control)  vs.  Challenger  (test)  A|B  Test  

Treatment  

Champion   Challenger  

Mailed   60850   52812  

Responses   1055   455  

Response  Rate   1.7   0.9  

Total  Value   $36,925   $38,675  

Mean  Value   $35   $85  

Std  Dev   $30   $50  

85−35±1.96* 502

455+302

105550± 4.9

At  a  minimum,  we  should  expect  an  incremental  $45.1  if  we  rolled  out  the  Challenger  crea0ve  as  BAU  (although  our  total  amount  of  incremental  revenue  would  be  less).  

Page 66: Testing for Success

>  Case  Study  

December  2011   ©  Datalicious  Pty  Ltd   66  

Page 67: Testing for Success

>  Main  Effects  

December  2011   ©  Datalicious  Pty  Ltd   67  

Page 68: Testing for Success

>  Main  Effects  

December  2011   ©  Datalicious  Pty  Ltd   68  

Element   Results  

Headline   Call  to  Ac&on   Social  Proof   Visitors  

Tested   Conversions   Conversion  Rate  

Treatment  

1   H1   CTA1   SP1   1237   456   37%  

2   H1   CTA1   SP2   1456   345   24%  

3   H1   CTA2   SP1   1245   234   19%  

4   H1   CTA2   SP2   2123   432   20%  

5   H2   CTA1   SP1   1342   234   17%  

6   H2   CTA1   SP2   1102   123   11%  

7   H2   CTA2   SP1   1365   700   51%  

8   H2   CTA2   SP2   1243   643   52%  

Typical  Landing  Page  Test  

Treatment  #7  and  #8  were  the  clear  winners  and  It  looks  as  if  the  Headline  and  Call-­‐to-­‐Ac0on  were  much  bigger  drivers  of  posi0ve  performance  than  the  Social  Proof.  Lets  check  this!  

Page 69: Testing for Success

>  Main  Effects  

December  2011   ©  Datalicious  Pty  Ltd   69  

Typical  Landing  Page  Test  

Element   Results  

Headline   Call  to  Ac&on  

Social  Proof  

Visitors  Tested  

Conversion  Rate  

Treatment  

1   H1   CTA1   SP1   1237   37%  

2   H1   CTA1   SP2   1456   24%  

3   H1   CTA2   SP1   1245   19%  

4   H1   CTA2   SP2   2123   20%  

5   H2   CTA1   SP1   1342   17%  

6   H2   CTA1   SP2   1102   11%  

7   H2   CTA2   SP1   1365   51%  

8   H2   CTA2   SP2   1243   52%  

Avg  H1=24%  

The  Main  Effect  of  the  Headline  is  simply  the  (weighted)  average  conversion  rate  for  Headline  2  less  the  (weighted)  average  conversion  rate  for  Headline  1    (33%-­‐24%=9%)  

Avg  H2=33%  

Page 70: Testing for Success

>  Main  Effects  

December  2011   ©  Datalicious  Pty  Ltd   70  

Typical  Landing  Page  Test  

Main  Effect  

Element  Headline   9.4%  

Call  to  Ac&on   11.1%  Social  Proof   5.3%  

In  actual  fact,  it  was  varia0ons  in  Call  to  Ac0on  that  had  the  most  posi0ve  impact  on  our  results,  improving  conversions  by  11.1%  points.  

Page 71: Testing for Success

>  Interac&on  Effects  

December  2011   ©  Datalicious  Pty  Ltd   71  

Typical  Landing  Page  Test  

Element   Results  

Headline   Call  to  Ac&on  

Social  Proof  

Visitors  Tested  

Conversion  Rate  

Treatment  

1   H1   CTA1   SP1   1237   37%  

2   H1   CTA1   SP2   1456   24%  

7   H2   CTA2   SP1   1365   51%  

8   H2   CTA2   SP2   1243   52%  

3   H1   CTA2   SP1   1245   19%  

4   H1   CTA2   SP2   2123   20%  

5   H2   CTA1   SP1   1342   17%  

6   H2   CTA1   SP2   1102   11%  

An  interac0on  effect  is  present  where  the  performance  of  one  element  is  dependent  on  which  varia0on  of  the  another  variable  is  present.  In  this  example,  we  are  looking  at  whether  the  results  for  each  of  the    Headlines  is  dependent  on  which  Call-­‐to-­‐Ac0on.  

Page 72: Testing for Success

>  Interac&on  Effects  

December  2011   ©  Datalicious  Pty  Ltd   72  

Typical  Landing  Page  Test  

Element   Results  

Headline   Call  to  Ac&on  

Social  Proof  

Visitors  Tested  

Conversion  Rate  

Treatment  

1   H1   CTA1   SP1   1237   37%  

2   H1   CTA1   SP2   1456   24%  

3   H1   CTA2   SP1   1245   19%  

4   H1   CTA2   SP2   2123   20%  

5   H2   CTA1   SP1   1342   17%  

6   H2   CTA1   SP2   1102   11%  

7   H2   CTA2   SP1   1365   51%  

8   H2   CTA2   SP2   1243   52%  

Wtd  Avg  H1CTA1=30%  

The  first  step  is  to  create  weighted  average  response  rates  between  for  each  of  the  two  factors  (ignoring  Social  Proof).    

Wtd  Avg  H1CTA2=20%  

Wtd  Avg  H2CTA1=14%  

Wtd  Avg  H2CTA2=51%  

Page 73: Testing for Success

>  Interac&on  Effects  

December  2011   ©  Datalicious  Pty  Ltd   73  

Typical  Landing  Page  Test  

Call  to  Ac&on  

CTA1   CTA2   Diff  

Headline  

H1   30%   20%   -­‐10%  

H2   14%   51%   37%  

Diff   -­‐16%   31%  

The  next  step  is  to  calculate  the  difference  in  performance  of  one  factor  across  different  variants  of  the  other  factor.  If  the  difference  of  this  difference  is  non-­‐zero  (or  not  very  close  to  zero),  then  you  have  an  interac0on  effect.      For  example,  there  is  an  interac0on  effect  between  the  Headline  and  Call  to  Ac0on  as  the  difference  in  the  difference  in  performance  is  non-­‐zero  (31%-­‐(-­‐16%)=47%).  This  is  very  large  interac0on  when  compared  to  the  Main  Effects!  

0%  

20%  

40%  

60%  

H1   H2  

CTA1  

CTA2  

Page 74: Testing for Success

>  Interac&on  Effects  

December  2011   ©  Datalicious  Pty  Ltd   74  

Typical  Landing  Page  Test  

Sociol  Proof  

SP1   SP2   Diff  

Headline  

H1   28%   22%   -­‐6%  

H2   34%   33%   -­‐1%  

Diff   -­‐6%   11%  0%  

20%  

40%  

H1   H2  

SP1  

SP2  

Sociol  Proof  

SP1   SP2   Diff  

Call  to  Ac&on  

CTA1   27%   18%   -­‐9%  

CTA2   36%   32%   -­‐4%  

Diff   9%   14%  

0%  

20%  

40%  

CTA1   CTA2  

SP1  

SP2  

Page 75: Testing for Success

>  Repor&ng  

101011010010010010101111010010010101010100001011111001010101010100101011001100010100101001101101001101001010100111001010010010101001001010010100100101001111101010100101001001001010  

December  2011   ©  Datalicious  Pty  Ltd   75  

Page 76: Testing for Success

December  2011   ©  Datalicious  Pty  Ltd   76  

Document  Everything!  

Page 77: Testing for Success

>  1.  Describe  the  test  

§  Describe  the  outcome(s)  you’re  trying  to  influence  

§  Describe  your  target  audience  §  Describe  the  different  treatments  including  copies  of  crea0ve  

December  2011   ©  Datalicious  Pty  Ltd   77  

Page 78: Testing for Success

>  2.  Jus&fy  the  test  design  

§  Detail  why  you’ve  chosen  the  par0cular    outcome  you’re  trying  to  influence  

§  Detail  why  you’ve  chosen  the  consumers  you  are  trying  to  influence  

§  Detail  why  your  interven0on  should  work  – Past  test  results/Useability  test/Case  studies  – Marketers  intui0on/logic  

December  2011   ©  Datalicious  Pty  Ltd   78  

Page 79: Testing for Success

>  3.  Results  &  Conclusions  

§  Detail  all  the  performance  results  §  Discuss  your  hypotheses  §  Future  tests  §  ‘Meta’  repor0ng  of  your  test  program  

 

December  2011   ©  Datalicious  Pty  Ltd   79  

Page 80: Testing for Success

>  The  Scien&fic  Method  

©  Datalicious  Pty  Ltd   80  

Knowledge  

Data  

Develop  Test(s)  

Establish  Facts  

December  2011  

Page 81: Testing for Success

>  Case  Study  

December  2011   ©  Datalicious  Pty  Ltd   81  

Page 82: Testing for Success

>  List  of  (Some)  Resources  

§  hRp://visualwebsiteop0mizer.com/case-­‐studies.php  

§  hRp://www.whichtestwon.com/  §  hRp://www.feng-­‐gui.com  §  hRp://www.smashingmagazine.com/2010/06/24/the-­‐ul0mate-­‐guide-­‐to-­‐a-­‐b-­‐tes0ng  

December  2011   ©  Datalicious  Pty  Ltd   82  

Page 83: Testing for Success

December  2011   ©  Datalicious  Pty  Ltd   83  

Contact  us  [email protected]  

 Learn  more  

blog.datalicious.com    

Follow  us  twizer.com/datalicious  

 

Page 84: Testing for Success

Data  >  Insights  >  Ac&on  


Top Related