Skip to content

LtvnSergey/Advertising-A-B-Testing

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Advertising-A-B-testing

Project overview

The company is trying new ad design to increase ad success. To do so they made new creative advertising design with SmartAd brand and conduct an A/B test in which one group of users have been shown an old ad and another - new one.

Whole analysis and results can be find in the following notebook.

Contents

Project structure

Data description

  • Input data - csv file with following columns:
Field Description
auction_id The unique id of the online user who has been presented the BIO. In standard terminologies this is called an impression id. The user may see the BIO questionnaire but choose not to respond. In that case both the yes and no columns are zero.
experiment Which group the user belongs to - control or exposed
control Users who have been shown a dummy ad
exposed Users who have been shown a creative, an online interactive ad, with the SmartAd brand
date The date in YYYY-MM-DD format
hour The hour of the day in HH format
device_make The name of the type of device the user has e.g. Samsung
platform_os The id of the OS the user has
browser The name of the browser the user uses to see the BIO questionnaire
yes 1 if the user chooses the “Yes” radio button for the BIO questionnaire
no 1 if the user chooses the “No” radio button for the BIO questionnaire

Experiment approach

Our goal is to analyse the results of A/B test and figure out wherever new disign of ad affects the behavior of users in terms of responding to BIO questionnaire.

  • Null Hypothesis Hₒ: p = pₒ - There is no significant difference between the ad success rate of both groups

  • Alternative Hypothesis Hₐ: p ≠ pₒ - There is significant difference between the ad success rate of both groups. Given we don’t know if the new design will perform better/worse/equal as our current design, we will perform a two-tailed test

  • Confidence Level: 95% (α=0.05)

  • p and pₒ stand for the conversion rate of the new and old design.

Conducted tests

1. Fisher exact test

  • Fisher exact test was used to show how response-rate variable differs between control and exposed group

  • Based on contingency table p-value was calculated

image

  • Result: Fisher test p-value: 0.53, therefor we cannot reject Hₒ hypothesis

2. Confidence Interval

  • Calculating bounds of a confidence interval, within which absolute difference between mean of the two group suppose to be.

image

  • Result: Confidence interval showed us that with 95% difference between two mean can be found between [-0.04, 0.07] which means we cannot reject Null hypothesis.

Conclusion

  • Based on the results of conducted A/B test with Fisher exact test and confidence interval we can conclude that new ad design havent changed the conversion rate.

Modules and tools

Python | Pandas | Numpy | Stats