The new forums will be named Coin Return (based on the most recent
vote)! You can check on the status and timeline of the transition to the new forums
here.
The Guiding Principles and New Rules
document is now in effect.
Uncertainty and Weighted Averages (Statistics)
Hi all, it's been a while since I've run one of these and I just want to get some verification that I'm approaching this the right way before I finalize and move on.
I have measurements of a quantity (specifically, a quantity in a beaker); three independent trials with 5 measurements each. Thus far, I've computed the individual mean and standard error for each trial (giving me three estimates of mean ± SE). I want to propagate the error of this through a system, so I'm going to want the overall best estimate of the measurement for this beaker quantity. In order to do that, what I've done at the moment is taken the weighted average using each mean and each uncertainty (which itself requires a small error propagation for the final uncertainty term) to calculate the overall "best estimate" of the beaker quantity. Given that I have three independent trials this seemed to make the most logical sense (rather than, say, taking the grand mean of all 15 replicates across the three trials and calculating the SE from that). Could I bother someone to verify that this is probably the best/most precise method for doing this? Any input is much appreciated. Feel free to ask if I've left any pertinent information out.
0
Posts
More generally: why not aggregate your trials? If they really are independent and measuring the same thing, this will tend to give better answers for most methods than arbitrarily dividing into three groups. As I mentioned above, this will give a smaller variance, which is generally what you want.
I don't understand what you mean by `propagating through a system'. Does this mean that you have some direct measurements (e.g. beaker volume) and would like to apply a function to the measurements (e.g. amount of time it takes for some Rube-Goldberg machine to run when the beaker is poured into a funnel)? If so, there are a few ways to deal with it. If you read a standard undergrad textbook, they will tell you to use things like `the delta method', otherwise known as taking a Taylor expansion of the function. Since you only have 15 data points, it is probably easiest to just apply your function to those points directly. There are a few cases where this won't do as well (e.g. if your function is very spiky AND you known a great deal about your function), so write again if you have some reason for suspecting this. If you're worried about these situations, and computing the function is easy, a first remedy might be to try to generate many (e.g. 10,000) data points by adding little bits of fuzz to your data and see if the resulting distribution is obviously screwy relative to that of the original points.
I hope this helps a little!
Regarding propagation: I'm actually referring more generally to a propagation of error. That is, this measured beaker quantity is then used in subsequent calculations, and I'm propagating the uncertainty (based upon these empirical measurements) throughout said system using the best estimate of the quantity (its mean) and its uncertainty that I can calculate.
Hopefully I hit upon your points; please let me know if I neglected anything! And thanks again for the response; it's very much appreciated.
P2: I meant exactly the same thing, sorry for being vague. I'm glad this is the case. So, again, it seems reasonable to do the calculations for each measurement separately if you're just interested in the final values, and then calculate summary statistics at the end - in other words, only compute the mean and variance of the thing you actually want. This can get substantially better answers than e.g. the Delta method. The fuzzing I mentioned was just a way to check against some possibly hidden numerical instability in doing this. As is often the case in statistics, there is a whole theory of how to propagate error through functions... but it gets pretty mathy pretty quickly.