TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

How Optimizely Almost Got Me Fired (2014)

163 pointsby yanowitzover 9 years ago

24 comments

iamleppertover 9 years ago
A&#x2F;B testing is high on hype and promise but low on actual results if you follow it though to actual metrics. I&#x27;ve done various forms of A&#x2F;B throughout most of my career and found them to be cinsistent with the OP&#x27;s results.<p>A much better approach is to install significant instrumentation and actually talk to users about what&#x27;s wrong with your sign up form.<p>That, or actually build a product that users want instead of chasing after pointless metrics. I mean, really, you think changing the color of text or a call-out is going to make up for huge deficiencies in your product or make people buy it? The entire premise seems illogical and just doesn&#x27;t work. The only time I&#x27;ve seen a&#x2F;b tests truly help was when it accidentally fixed some cross browser issue or moved a button within reach of a user.<p>Most of the A&#x2F;B website optimization industry is an elaborate scam, put off on people who don&#x27;t know any better and are looking for a magic bullet.
评论 #10874269 未加载
评论 #10873538 未加载
评论 #10874754 未加载
评论 #10873410 未加载
评论 #10873330 未加载
评论 #10873174 未加载
评论 #10873775 未加载
评论 #10873316 未加载
评论 #10873329 未加载
评论 #10873948 未加载
评论 #10874691 未加载
评论 #10873164 未加载
评论 #10873462 未加载
feralover 9 years ago
[I&#x27;m a PM @ Optimizely]<p>We were asked about this article before, on our community forums, and one of our statisticians, David, wrote a detailed reply to this article&#x27;s concerns about one- vs two-tailed testing, which might be of interest [3rd from the top]:<p><a href="https:&#x2F;&#x2F;community.optimizely.com&#x2F;t5&#x2F;Strategy-Culture&#x2F;Let-s-talk-about-Single-Tailed-vs-Double-Tailed&#x2F;m-p&#x2F;4220" rel="nofollow">https:&#x2F;&#x2F;community.optimizely.com&#x2F;t5&#x2F;Strategy-Culture&#x2F;Let-s-t...</a><p>Additionally, since then, as other commenters have mentioned, we&#x27;ve completely overhauled how we do our A&#x2F;B-testing calculations, which, theoretically and empirically, now have an accurate false-positive rate even when monitored continuously. Details:<p><a href="https:&#x2F;&#x2F;blog.optimizely.com&#x2F;2015&#x2F;01&#x2F;20&#x2F;statistics-for-the-internet-age-the-story-behind-optimizelys-new-stats-engine&#x2F;" rel="nofollow">https:&#x2F;&#x2F;blog.optimizely.com&#x2F;2015&#x2F;01&#x2F;20&#x2F;statistics-for-the-in...</a>
yummyfajitasover 9 years ago
Disclaimer: I&#x27;m the director of data science at VWO, an Optimizely competitor.<p>In my view, the issue is not one-tail vs two-tail tests, or sequential vs one-look tests at all. The issue is a failure to quantify uncertainty.<p>Optimizely (last time I looked), our old reports, and most other tools, all give you improvement as a single number. Unfortunately that&#x27;s BS. It&#x27;s simply a lie to say &quot;Variation is 18% better than Control&quot; unless you had facebook levels oftraffic. An honest statement will quantify the uncertainty: &quot;Variation is between -4.5% and +36.4% better than Control&quot;.<p>When phrased this way, it&#x27;s hardly surprising that deploying this variation failed to achieve an 18% lift - 18% is just one possible value in a wide range of possible values.<p>The big problem with this is that customers (particularly agencies who are selling A&#x2F;B test results to clients) hate it. If we were VC funded, we might even have someone pushing us to tell customers the lie they want rather than the truth they need.<p>Note that to provide uncertainty bounds like this, one needs to use a Bayesian method (only us, AB Tasty and Qubit do this, unless I forgot about someone).<p>(Frequentist methods can provide confidence intervals, but these are NOT the same thing. Unfortunately p-values and confidence intervals are completely unsuitable for reporting to non-statisticians; they are completely misinterpreted by almost 100% of laypeople. <a href="http:&#x2F;&#x2F;myweb.brooklyn.liu.edu&#x2F;cortiz&#x2F;PDF%20Files&#x2F;Misinterpretations%20of%20Significance.pdf" rel="nofollow">http:&#x2F;&#x2F;myweb.brooklyn.liu.edu&#x2F;cortiz&#x2F;PDF%20Files&#x2F;Misinterpre...</a> <a href="http:&#x2F;&#x2F;www.ejwagenmakers.com&#x2F;inpress&#x2F;HoekstraEtAlPBR.pdf" rel="nofollow">http:&#x2F;&#x2F;www.ejwagenmakers.com&#x2F;inpress&#x2F;HoekstraEtAlPBR.pdf</a> )
评论 #10874001 未加载
评论 #10875020 未加载
scuba_man_spiffover 9 years ago
One thing I noticed that I haven&#x27;t seen commented on yet:<p>The solution mentioned of running a two tailed test would not have solved the problem of a false result the author demonstrated through conducting an A&#x2F;A test.<p>According to the image in the article: <a href="http:&#x2F;&#x2F;blog.sumall.com&#x2F;wp-content&#x2F;uploads&#x2F;2014&#x2F;06&#x2F;optimizely-test.png" rel="nofollow">http:&#x2F;&#x2F;blog.sumall.com&#x2F;wp-content&#x2F;uploads&#x2F;2014&#x2F;06&#x2F;optimizely...</a><p>The A&#x2F;A test had: A1: Population: 3920 Conversion: 721 A2: Population: 3999 Conversion: 623<p><pre><code> Z-Score: 3.3 2-tailed test signifiance: 99.92% </code></pre> Looks like the one-tail vs. two tail test doesn&#x27;t make huge difference in this case.<p>So, maybe a larger sample size would have seen a reversion to the mean, but given the size and high significance that would be unlikely (interesting exercise to try different assumptions to calculate how unlikely, with the most overly generous obviously just being the stated significance).<p>Yes, the test was only conducted over one day, but if it was the exact same thing being served for both, that shouldn&#x27;t matter.<p>If there was a reversion to the mean due to an early spike, we would expect to see the % difference between the two cells narrow as the test kept running. You can see in the chart that the % difference (relative gap between the lines) stays about the same after 8pm on the 9th.<p>So if it&#x27;s not the one-tailed test at fault, and it&#x27;s not the short duration of the test at fault, what is?<p>Don&#x27;t know.<p>I have seen in the past that setup problems are incredibly easy to make w&#x2F; a&#x2F;b testing tools when implementing the tool on your site. I&#x27;ve seen in other tools things like automated traffic from Akamai only going to the default control, or subsets of traffic such as returning visitors excluded from some cells but not others.<p>Based on those results, I&#x27;d be suspicious of something in the tool setup being amiss.
closedover 9 years ago
&gt; This usually happens because someone runs a one-tailed test that ends up being overpowered.<p>It always pains me a little when people doing research describe statistical power as a type of curse. Overpowered? Should we reduce it? The risk isn&#x27;t having too much power, the risk is that someone will incorrectly interpret their Null Hypothesis Significance Test (NHST). They need to shift their focus to measuring something (and quantifying the uncertainty of their measurements), rather than think of &quot;how likely was this result given a null hypothesis&quot;, whether that hypothesis is..<p>something is not greater than 0 (one-tail), or<p>something is not 0 (two-tail).<p>&gt; You’ll often see statistical power conveyed as P90 or 90%. In other words, if there’s a 90% chance A is better than B, there’s a 10% chance B is better than A and you’ll actually get worse results.<p>This isn&#x27;t necessarily true. A could be the same as B. Also, these tests are being done from the frequentist perspective, so saying &quot;there&#x27;s X chance B is better than A&quot; is inappropriate, unless you&#x27;re talking about the conclusions of your significance test (e.g. 90% chance you correctly detect a difference between them--a difference you assume is fixed to some true underlying value). Overall, being aware that a one-tail test is taking the position that nothing can happen in the other direction is useful, but a good next step is understanding what NHST can and cannot say.<p>This even a frequentist vs bayesian problem, since you could create situations where a person felt a study was overpowered in either framework.
jacalataover 9 years ago
Don&#x27;t just run tests longer - run tests for a pre-defined amount of time instead of &quot;until you see a result you like&quot;.
评论 #10874015 未加载
评论 #10873415 未加载
评论 #10874886 未加载
ellipticover 9 years ago
I agree with the main point of the article, but I&#x27;m somewhat disturbed by the statistical errors and misconceptions.<p>&gt;Few websites actually get enough traffic for their &gt;audiences to even out into a nice pretty bell curve. If &gt;you get less than a million visitors a month your &gt;audience won’t be identically distributed and, even &gt;then, it can be unlikely.<p>What is the author trying to say here? Has he thought hard about what it means for &quot;an audience&quot; to be identically distributed?<p>&gt;Likewise, the things that matter to you on your website, &gt;like order values, are not normally distributed Why do they need to be?<p>&gt;Statistical power is simply the likelihood that the &gt;difference you’ve detected during your experiment &gt;actually reflects a difference in the real world. Simply googling the term would reveal this is incorrect.
hb42over 9 years ago
&gt; In most organizations, if someone wants to make a change to the website, they’ll want data to support that change.<p>So true and sad. In all the so called data-driven groups I have worked for, the tyranny of data makes metrics and numbers the justification for or counter to anything, however they have been put together.<p>&gt; The sad truth is that most people aren’t being rigorous about their A&#x2F;B testing and, in fact, one could argue that they’re not A&#x2F;B testing at all, they’re just confirming their own hypotheses.<p>The sad truth is that most people aren’t being rigorous about anything.
RA_Fisherover 9 years ago
Here&#x27;s a great article about how Optimizely gets it wrong: <a href="http:&#x2F;&#x2F;dataorigami.net&#x2F;blogs&#x2F;napkin-folding&#x2F;17543303-the-binary-problem-and-the-continuous-problem-in-a-b-testing" rel="nofollow">http:&#x2F;&#x2F;dataorigami.net&#x2F;blogs&#x2F;napkin-folding&#x2F;17543303-the-bin...</a><p>There are _many_ offenders. I&#x27;ve yet to see a commercial tool that gets it right.<p>Tragically, the revamp by Optimizely neglects the straightforward Bayesian solution and uses a more fragile and complex sequential technique.
评论 #10873136 未加载
kylerushover 9 years ago
Optimizely rolled out a huge update to the way it handles statistics called Stats Engine last year. That update resolves the issues discussed in this article. You can read more about Stats Engine here: <a href="https:&#x2F;&#x2F;www.optimizely.com&#x2F;statistics&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.optimizely.com&#x2F;statistics&#x2F;</a>
评论 #10873400 未加载
cwyersover 9 years ago
&gt; some testing vendors use one-tailed tests (Visual Website Optimizer, Optimizely)<p>&gt; Most A&#x2F;B testing tools recommend terminating tests as soon as they show significance, even though that significance may very well be due to short-term bias. A little green indicator will pop up, as it does in Optimizely, and the marketer will turn the test off.<p>People pay brisk money for this?
stdbrouwover 9 years ago
After reading the blog post and reading through the comments, it looks like people are drawing the wrong conclusion from this. The problem is not that AB-testing is overrated, doesn&#x27;t work, is bullshit etc. but that Optimizely used to do it wrong.
评论 #10873809 未加载
评论 #10874678 未加载
erikbernover 9 years ago
&gt; Statistical power is simply the likelihood that the difference you’ve detected during your experiment actually reflects a difference in the real world.<p>This seems incorrect to me. Isn&#x27;t statistical power the likelihood that the null hypothesis would generate an outcome at least as extreme as what you observed?<p>I&#x27;m guessing the issue has a lot more to do with peeking at the outcome and not correcting for it (and similarly running many tests)<p><a href="http:&#x2F;&#x2F;www.stat.columbia.edu&#x2F;~gelman&#x2F;research&#x2F;unpublished&#x2F;p_hacking.pdf" rel="nofollow">http:&#x2F;&#x2F;www.stat.columbia.edu&#x2F;~gelman&#x2F;research&#x2F;unpublished&#x2F;p_...</a>
评论 #10873170 未加载
评论 #10874566 未加载
nartzover 9 years ago
In my experience, if you are seeing huge effects like 60% difference in conversion, etc, you probably did something wrong (i.e. too small sample, didn&#x27;t wait long enough, etc) - I&#x27;ve never seen something this large by simply moving things around, changing colors, changing messages, etc.<p>Also, in general, the more drastic the changes are, the more of an effect you <i></i>could<i></i> have (up to some percentage). I.e. a small change would be changing a message or color, dont expect conversion to change by much. A large change would be changing from a flash site to an html one with a full redesign that loads twice as fast...
评论 #10875162 未加载
forrestthewoodsover 9 years ago
I&#x27;m increasingly convinced all statistical analysis performed by non-PhDs is no better than a coin flip. Maybe even worse.<p>My favorite example is still the quite popular Page Weight Matter posts. I wonder how close they were to abandoning a 90% reduction in size. I wonder how many improvements the world at large has thrown away due to faulty analysis.<p><a href="http:&#x2F;&#x2F;blog.chriszacharias.com&#x2F;page-weight-matters" rel="nofollow">http:&#x2F;&#x2F;blog.chriszacharias.com&#x2F;page-weight-matters</a>
jamiequintover 9 years ago
This is a huge problem with paid marketing as well. Many folks will look at a conversion rate completely ignorant of sample size and allocate thousands of dollars in budget to something which they have no idea performs better or worse.<p>The real problem (as you allude to in the article) is that the demand for accurate tools is not really there. Vendors don&#x27;t build in accurate stats because only a tiny portion of their client base understand&#x2F;demands them.
anarchitectover 9 years ago
There is much more to running experiments properly than it seems. While I&#x27;m not an expert on the statistics side, there are a few things I&#x27;ve learned over the years which come to mind...<p>1) Run the experiment in whole business cycles (for us, 1 week = one cycle), based on a sample size you&#x27;ve calculated upfront (I use <a href="http:&#x2F;&#x2F;www.evanmiller.org&#x2F;ab-testing&#x2F;sample-size.html" rel="nofollow">http:&#x2F;&#x2F;www.evanmiller.org&#x2F;ab-testing&#x2F;sample-size.html</a>). Accept that some changes are just not testable in any sensible amount of time (I wonder what the effect of changing a font will have on e-commerce conversion rate).<p>2) Use more than one set of metrics for analysis to discover unexpected effects. We use the Optimizely results screen for general steer, but do final analysis in either Google Analytics or our own databases. Sometimes tests can positively affect the primary metric but negatively affect another.<p>3) Get qualitative feedback either before or during the test. We use a combination of user testing (remote or moderated) and session recording (we use Hotjar, and send tags so we can view sessions in that experiment).
filleokusover 9 years ago
Interesting read, might be worth adding (2014) to the title though.
评论 #10873119 未加载
gnicholasover 9 years ago
Almost didn&#x27;t click on this because the title seems (and is) clickbait-y, but it was actually a very useful read for me.<p>As a founder, I&#x27;m constantly hearing about A&#x2F;B testing and how great these tools are. I&#x27;m not enough of a statistician to know whether everything in this article is true&#x2F;valid (and would welcome a rebuttal), but the part about regression to the mean really hits home. Encouraging users to cut off testing too early means that you make them feel good (&quot;Look, we had this huge difference!&quot;), when in reality the difference is smaller&#x2F;negligible.<p>I&#x27;ll still do some A&#x2F;B testing, but given our engineering&#x2F;time constraints—and my inability to accurately vet the claims&#x2F;conclusions of the testing software—I won&#x27;t spend too much time on this.
评论 #10876386 未加载
IndianAstronautover 9 years ago
AB testing is too simplistic. Even on my marketing team we have designed more complex metrics to look at a factors impact on multiple outcomes. The testing is still a straight forward chi square, but with a bit more depth.
TeMPOraLover 9 years ago
Wow, did not see that coming. This article actually confirms the cynical hypothesis I entertain - that most of the &quot;data-driven&quot; marketing and analytics is basically marketers bullshitting each other, their bosses, their customers and themselves, because nobody knows much statistics and everyone wants to believe that if they&#x27;re spending money and doing something, it must be bringing results.<p>Some quotes from the article supporting the cynical worldview:<p>--<p>&quot;Most A&#x2F;B testing tools recommend terminating tests as soon as they show significance, even though that significance may very well be due to short-term bias. A little green indicator will pop up, as it does in Optimizely, and the marketer will turn the test off. But most tests should run longer and in many cases it’s likely that the results would be less impressive if they did. Again, this is a great example of the default settings in these platforms being used to increase excitement and keep the users coming back for more.&quot;<p>This basically stops short of implying that Optimizely is doing this totally on purpose.<p>--<p>&quot;In most organizations, if someone wants to make a change to the website, they’ll want data to support that change. Instead of going into their experiments being open to the unexpected, open to being wrong, open to being surprised, they’re actively rooting for one of the variations. Illusory results don’t matter as long as they have fodder for the next meeting with their boss. And since most organizations aren’t tracking the results of their winning A&#x2F;B tests against the bottom line, no one notices.&quot;<p>In other words, everybody is bullshitting everybody, but it doesn&#x27;t matter as long as everyone plays along and money keeps flowing.<p>--<p>&quot;Over the years, I’ve spoken to a lot of marketers about A&#x2F;B testing and conversion optimization, and, if one thing has become clear, it’s how unconcerned with statistics most marketers are. Remarkably few marketers understand statistics, sample size, or what it takes to run a valid A&#x2F;B test.&quot;<p>&quot;Companies that provide conversion testing know this. Many of those vendors are more than happy to provide an interface with a simple mechanic that tells the user if a test has been won or lost, and some numeric value indicating by how much. These aren’t unbiased experiments; they’re a way of providing a fast report with great looking results that are ideal for a PowerPoint presentation. <i>Most conversion testing is a marketing toy, essentially</i>.&quot; (emphasis mine)<p><i>Thank you</i> for admitting it publicly.<p>--<p>Like whales, whose cancers grow so big that the tumors catch their own cancers and die[0], it seems that marketing industry, a well known paragon of honesty and teacher of truth, is actually being held down by its own utility makers applying their honourable strategies within their own industry.<p>I know it&#x27;s not a very appropriate thing to do, but I <i>really</i> want to laugh out loud at this. Karma is a bitch. :).<p>[0] - <a href="http:&#x2F;&#x2F;www.nature.com&#x2F;news&#x2F;2007&#x2F;070730&#x2F;full&#x2F;news070730-3.html" rel="nofollow">http:&#x2F;&#x2F;www.nature.com&#x2F;news&#x2F;2007&#x2F;070730&#x2F;full&#x2F;news070730-3.htm...</a>
评论 #10874707 未加载
jmountover 9 years ago
(as others have mentioned) Optimizely&#x27;s newer engine uses ideas like Wald&#x27;s sequential analysis. Here is my article on the topic: <a href="http:&#x2F;&#x2F;www.win-vector.com&#x2F;blog&#x2F;2015&#x2F;12&#x2F;walds-sequential-analysis-technique&#x2F;" rel="nofollow">http:&#x2F;&#x2F;www.win-vector.com&#x2F;blog&#x2F;2015&#x2F;12&#x2F;walds-sequential-anal...</a> .
hyperpalliumover 9 years ago
\tangent When I first heard of A&#x2F;B testing, I thought of combining it with genetic algorithms to evolve the entire site. Just run it til the money rolls in.<p>Unfortunately, if it did work, it would probably be through something misleading or scammy. Therefore, you need some kind of automatic legality checking... which would be hard.
jbpetersenover 9 years ago
Is anybody out there taking an approach of gradually driving more traffic to whichever option is winning out and never running 100% with anything specific?