3) The last two sentences actually interpret the estimate of 1% as if it were the (unknown)
effect. This is dangerous even when the estimate IS statistically significant, since we still
only have an estimate of the effect and not the real thing.
What needs to be done differently? Well, we might need to learn to write things like:
“We are not able to identify the sign of the effect of the program on job placement.”
So sad. Too sad for most. (And too unpublishable?) So we try things like:
“The estimate is statistically insignificant, but it is positive as predicted by theory.”
Well…technically, “positive” is modifying “estimate,” so you are not making a false statement
here. Similarly you could use the word “coefficient” instead of “estimate.” But is this tiptoeing
around really the best way?
I am convinced that if we properly use confidence intervals, we will do a much better job of
avoiding accidental forays into overstating our findings (because, after all, that is what is being
done in these examples and we ought to admit it).
“Using our estimates, we cannot reject a zero effect, and our findings suggest that the range
from -3% to +5% likely contains the true effect.”
Technically, there is still the issue that we should test against different nulls (such as -2%, +4%,
or others) if we really want to say more about anything but zero (our null hypothesis). I leave
further discussion to someone even more in the statistical grammar weeds than me.
4. We just don’t know the true effect and that’s okay
Perhaps the most painful realization for someone trying to write accurately about statistical
analysis is that no matter what we do, we will not know if we have recovered the true beta. In
fact, we probably will not have recovered it (it’s basically a probability zero event, right?). The
thing we have will always be betahat. Now, betahat is very useful and can give us a lot of
insight, but many times we pretend that it is beta. Suppose, for the sake of argument, that
betahat is 0.05 with a standard error of 0.02. Likely write-ups might be things like:
“The job training program increased job placements by 5%.”
“There was a statistically-significant 5% effect of the training program on job placements.”
Unfortunately – along with the misplaced “statistically-significant” modifier of the word “effect”
in the second one – we are treating our estimate as if we have recovered beta. Actually,
though, we have a confidence interval of roughly .01 to .09 (i.e. 1% to 9%). Try these:
“We emerge confident that the job training program increased placements, and our estimates
suggest the improvement was likely between 1% and 9%.”
4