đ° Summary (use your own words)
- Frequentist approach to statistical tests significance can be prone to large sample size
âď¸ Notes
- Donât fall prey to ârepeated significance testing errorsâ
- When a A/B test says there is a â95% of chance of beating the originalâ or â90% probability of statistical significanceâ
- It is asking the question âwhat is the chance that we observe a difference like what we see in data randomly?â
- And that is the significance threshold (0.05 or 0.01)
- However, this is precedent on a key decision -> the sample size was fixed in advance
- Because might be tempted to take action once we see significant results and if that becomes insignificant later we wouldnât have known
- So the key is to determine a sample size before the experiement
- One rule of thumb is $N=16*\frac{\sigma^2}{\delta^2}$ where $\delta$ is the minimum effect you want to observe
- For Bayesian A/B testing, this is less of a problem compared to the frequentist approach. But not completely fool-proof -> Is Bayesian AB testing immune to peeking