

Study with the several resources on Docsity
Earn points by helping other students or get them with a premium plan
Prepare for your exams
Study with the several resources on Docsity
Earn points to download
Earn points by helping other students or get them with a premium plan
Community
Ask the community for help and clear up your study doubts
Discover the best universities in your country according to Docsity users
Free resources
Download our free guides on studying techniques, anxiety management strategies, and thesis advice from Docsity tutors
The importance of population variance in statistical inference, particularly in the context of risk assessment. It covers methods for estimating population variance, performing tests on variances, and testing for equality of variances between two populations. The document also introduces the chi-square and f distributions, which are used in these tests.
Typology: Exams
1 / 2
This page cannot be seen from the preview
Don't miss anything!
Inference (hypothesis tests and estimation) Variance of one population, comparing two variances
Tests and estimation on means and proportions focus on “typical value” or “where is the center of the values”. The other huge aspect of distributions (that is still simpler than “shape”) is variability. In dealing with (for example) the stock market and/or particular stocks, the interest is not just in “expected value” but in “risk” - the amount by which the value fluctuates. The most common measurement used is standard deviation and (its theoretically-nicer-but conceptually-wierder-brother) variance. We would like to be able to estimate a population standard deviation, to perform tests on a standard deviation, and to test for equality (or inequality) of standard deviations from two populations - but the work is carried out in terms of variance. Our best estimator for σ^2 is s^2 , if we look at all samples of size n, the mean of all the s^2 values will be exactly σ^2. [Because of the behavior of square roots, the values of s would not average out to the value of σ]. Distribution: The distribution of s^2 values is not normal — we can’t convert to a Z or t. If the variable (X) being observed is approximately normal, then the distribution of values of s^2 does fit the χ^2 (chi-square) family. We tend to get a χ^2 distribution if we add the squares of a lot of normal variables (as we do, in calculating s^2 , if X is normal). As with the t distributions, there is a different version of the χ^2 for each value of “‘degrees of freedom”; the distribution is not symmetric, and the mean gets larger with larger degrees of freedom. Table: We will work with a table of critical values for χ^2 — it’s a little more complicated than the t table because all values are positive (no ±tα tricks can be used) and the distribution is not symmetric. The value which cuts off area α on the right (larger) side is called χ^2 α (same as with Z or t) but the value that cuts off area α on the left can’t be −χ^2 α - we notice that it must cut off area 1 − α to the em right and call it χ^2 (1−α)
Test Statistic: To work with s^2 we use a normalization — but instead of subtracting the mean of the distribution, we divide by the mean (of the s^2 values — that is σ^2 of the X values) and multiply by degrees of freedom - so our test
statistic will be sample χ^2 = (n−1)s
2 σ^2.
Estimation We do not get an estimate of the form (statistic) ± E - we will be looking at multiplication/division: To estimate σ^2 , with confidence 1 − α we use the interval
(n − 1)s^2 χ^2 α/ 2
to (n − 1)s^2 χ^2 (1−α/2)
with df = n − 1
Tests For tests, we have the usual six steps, but the hypotheses are about σ^2 , the test statistic is sample χ^2 = (n−1)s
2 σ^2 and we have to look for χ^2 (1−α) or χ^2 (1−α/2) where we would have used −tα or −tα/ 2 for tests on a mean.
“Greater” H 0 : σ^2 = σ^20 Ha : σ^2 > σ 02
“Less” H 0 : σ^2 = σ^20 Ha : σ^2 < σ 02
“not equal” H 0 : σ^2 = σ^20 Ha : σ^2 6 = σ^20
sample χ^2 =
(n − 1)s^2 σ 02 df = n − 1
Reject H 0 if sample χ^2 > χ^2 α Reject H 0 if sample χ^2 < χ^2 (1−α) Reject H 0 if sample χ^2 < χ^2 (1− α 2 )
or sample χ^2 > χ^2 (1− α 2 )
Tests on difference of two variances To decide whether the variances of two populations are equal, we will consider the ratio of sample variances s
(^21) s^22.^ The distribution of this ratio will follow yet another distribution - this one for a ratio of χ^2 distributions. It’s called the F distribution and changes based on not one but two values - degrees of freedom for the numerator and degrees of freedom for the denominator. In this note I will use the notation F(1−α) ab for “the value of F , with degrees of freedom a for the numerator, v for the denominator, that cuts off area α to the right”. In your text (pp. 925-927) there is a table of critical values for the right-side only (that is Fα ab for small α). The conversion for the left-side value is not pretty but not terribly hard: The left-side critical value (for “ <′′^ and “ 6 =′′^ tests) is given by F(1−α) ab = (^) Fα^1 ba.
For the tests, we have the usual six steps, but the hypotheses are about σ 12 and σ^22 , the test statistic is sample F = s
(^21) s^22 and we have to look for F(1−α) ab or F(1−α/ 2 ) ab where we would have used −tα or −tα/ 2 for a test on means.
“Greater” H 0 : σ^21 = σ^22 Ha : σ 12 > σ 22
“Less” H 0 : σ^21 = σ^22 Ha : σ 12 < σ 22
“not equal” H 0 : σ^21 = σ^22 Ha : σ 12 6 = σ^22
sample F = s
(^21) s^22 numerator df = a = n 1 − 1 , denominator df = b = n 2 − 1
Reject H 0 if sample F > Fα ab Reject H 0 if sample F < F(1−α) ab Reject H 0 if sample F < F(1− α 2 ) ab or sample F > F(1− α 2 ) ab
Class examples for tests and estimation on variances.
(a) Does this give evidence that the standard deviation of prices is less than $10.00? (b) What is our 95% confidence estimate of the standard deviation of prices.
(a) Does this indicate at the .01 level that the variation in amounts is greater than specifications? (b) What is our 95% confidence estimate for the variance of the amounts?