I have a 40-long list of means and variances relating to model error rates. I am looking to determine which models are statistically better (smaller error rates) than others.
Assuming error rates are normally distributed, I am presently looking up z-scores manually, which is working but taking a long time. Is there a more pythonic way to create a matrix of probability scores comparing, for instance, model a vs model b, model b vs model c, model a vs model c?
I haven't included any code thus far as I am currently working in excel, but have included the dummy sample below if anyone can assist.
data = [[10, 0.8],[5, 1.2],[12, 2.4],[6, 2.8]]