Results 1 to 30 of 4194

Thread: What made you go WTF today?

Hybrid View

Previous Post Previous Post   Next Post Next Post
  1. #1
    Good catch on the journal name.

    If the purpose of the study is to see whether X has a statistically significant effect on Y, you can't just up the sample until your p-value reaches an acceptable level. This is p-hacking taken to the extreme, and the authors nonchalantly admit to it.

    If your data has substantial variability, then it should be hard to obtain statistical significance. That's the whole point. If you want to claim there's a large effect size, that's fine. If you want to claim hypothesis testing is unnecessary and that you're just providing summary statistics, that's more or less fine. But you can't have your cake and eat it too by doing hypothesis testing and then explicitly changing the parameters of the study until you get the desired result. Or at least you shouldn't admit to doing it.
    Hope is the denial of reality

  2. #2
    Quote Originally Posted by Loki View Post
    Good catch on the journal name.

    If the purpose of the study is to see whether X has a statistically significant effect on Y, you can't just up the sample until your p-value reaches an acceptable level. This is p-hacking taken to the extreme, and the authors nonchalantly admit to it.

    If your data has substantial variability, then it should be hard to obtain statistical significance. That's the whole point. If you want to claim there's a large effect size, that's fine. If you want to claim hypothesis testing is unnecessary and that you're just providing summary statistics, that's more or less fine. But you can't have your cake and eat it too by doing hypothesis testing and then explicitly changing the parameters of the study until you get the desired result. Or at least you shouldn't admit to doing it.
    I suspect that your critique is fair but I think it's important to distinguish this from p-hacking. P-hacking is essentially going in with no hypothesis, getting a shitload of data, and then running every test you can imagine until something 'significant' appears. This is distinct; they had a hypothesis but didn't know either the effect size or the population variance, so they scaled sample size to find the effect. It's potentially got some of the same problems with p-hacking if they didn't account for multiple post-hoc tests when reporting p-values (I suspect this is the case) - but it's not as dishonest as p-hacking. They're essentially saying 'we're pretty sure there's a signal here, the size of the effect we saw was X' rather than 'we had a hypothesis that Y effect exists and is this size, and sure enough we were right'. It's a bit sloppy science, since ideally they would have run a pilot study to inform their power analysis, but the fact of the matter is that no one does that.

    For subtle signals in in vivo data, I would indeed prefer to see at least one replication (or partial replication) before I'd submit this kind of data. And that would obviate the need to justify their sample size in this manner. But it's still informative IMO, even if it means your conclusions are far weaker than they might be otherwise.
    "When I meet God, I am going to ask him two questions: Why relativity? And why turbulence? I really believe he will have an answer for the first." - Werner Heisenberg (maybe)

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •