Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

Inference: Hypothesis Tests and Estimation for Population Variance, Exams of Mathematics

The importance of population variance in statistical inference, particularly in the context of risk assessment. It covers methods for estimating population variance, performing tests on variances, and testing for equality of variances between two populations. The document also introduces the chi-square and f distributions, which are used in these tests.

Typology: Exams

Pre 2010

Uploaded on 08/05/2009

koofers-user-ez4
koofers-user-ez4 🇺🇸

10 documents

1 / 2

Toggle sidebar

This page cannot be seen from the preview

Don't miss anything!

bg1
Inference (hypothesis tests and estimation)
Variance of one population, comparing two variances
Tests and estimation on means and proportions focus on “typical value” or “where is the center of the values”.
The other huge aspect of distributions (that is still simpler than “shape”) is variability. In dealing with (for example)
the stock market and/or particular stocks, the interest is not just in “expected value” but in “risk” - the amount by
which the value fluctuates. The most common measurement used is standard deviation and (its theoretically-nicer-but
conceptually-wierder-brother) variance.
We would like to be able to estimate a population standard deviation, to perform tests on a standard deviation, and
to test for equality (or inequality) of standard deviations from two populations - but the work is carried out in terms of
variance. Our best estimator for σ2is s2, if we look at all samples of size n, the mean of all the s2values will be exactly
σ2. [Because of the behavior of square roots, the values of swould not average out to the value of σ].
Distribution: The distribution of s2values is not normal we can’t convert to a Zor t. If the variable (X) being
observed is approximately normal, then the distribution of values of s2does fit the χ2(chi-square) family. We tend to
get a χ2distribution if we add the squares of a lot of normal variables (as we do, in calculating s2, if X is normal). As
with the tdistributions, there is a different version of the χ2for each value of “‘degrees of freedom”; the distribution is
not symmetric, and the mean gets larger with larger degrees of freedom.
Table: We will work with a table of critical values for χ2 it’s a little more complicated than the ttable because all
values are positive (no ±tαtricks can be used) and the distribution is not symmetric. The value which cuts off area αon
the right (larger) side is called χ2
α(same as with Zor t) but the value that cuts off area αon the left can’t be χ2
α- we
notice that it must cut off area 1 αto the em right and call it χ2
(1α)
Test Statistic: To work with s2we use a normalization but instead of subtracting the mean of the distribution,
we divide by the mean (of the s2values that is σ2of the X values) and multiply by degrees of freedom - so our test
statistic will be sample χ2=(n1)s2
σ2.
Estimation
We do not get an estimate of the form (statistic) ±E - we will be looking at multiplication/division:
To estimate σ2, with confidence 1 αwe use the interval
(n1)s2
χ2
α/2
to (n1)s2
χ2
(1α/2)
with df =n1
Tests
For tests, we have the usual six steps, but the hypotheses are about σ2, the test statistic is sample χ2=(n1)s2
σ2and we
have to look for χ2
(1α)or χ2
(1α/2) where we would have used tαor tα/2for tests on a mean.
“Greater”
H0:σ2=σ2
0
Ha:σ2> σ2
0
“Less”
H0:σ2=σ2
0
Ha:σ2< σ2
0
“not equal”
H0:σ2=σ2
0
Ha:σ26=σ2
0
sample χ2=(n1)s2
σ2
0
df =n1
Reject H0if sample χ2> χ2
αReject H0if sample χ2< χ2
(1α)Reject H0if sample χ2< χ2
(1α
2)
or sample χ2> χ2
(1α
2)
Tests on difference of two variances
To decide whether the variances of two populations are equal, we will consider the ratio of sample variances s2
1
s2
2
. The
distribution of this ratio will follow yet another distribution - this one for a ratio of χ2distributions. It’s called the F
distribution and changes based on not one but two values - degrees of freedom for the numerator and degrees of freedom
for the denominator. In this note I will use the notation F(1α)a
bfor “the value of F, with degrees of freedom afor the
numerator, vfor the denominator, that cuts off area αto the right”. In your text (pp. 925-927) there is a table of critical
values for the right-side only (that is Fαa
bfor small α). The conversion for the left-side value is not pretty but not terribly
hard: The left-side critical value (for <00 and 6=00 tests) is given by F(1α)a
b=1
Fαb
a.
For the tests, we have the usual six steps, but the hypotheses are about σ2
1and σ2
2, the test statistic is sample F=s2
1
s2
2
and we have to look for F(1α)abor F(1α/2)abwhere we would have used tαor tα/2for a test on means.
“Greater”
H0:σ2
1=σ2
2
Ha:σ2
1> σ2
2
“Less”
H0:σ2
1=σ2
2
Ha:σ2
1< σ2
2
“not equal”
H0:σ2
1=σ2
2
Ha:σ2
16=σ2
2
sample F=s2
1
s2
2
numerator df =a=n11,
denominator df =b=n21
Reject H0if sample F > Fαa
bReject H0if sample F < F(1α)a
bReject H0if sample F < F(1α
2)a
b
or sample F > F(1α
2)a
b
1
pf2

Partial preview of the text

Download Inference: Hypothesis Tests and Estimation for Population Variance and more Exams Mathematics in PDF only on Docsity!

Inference (hypothesis tests and estimation) Variance of one population, comparing two variances

Tests and estimation on means and proportions focus on “typical value” or “where is the center of the values”. The other huge aspect of distributions (that is still simpler than “shape”) is variability. In dealing with (for example) the stock market and/or particular stocks, the interest is not just in “expected value” but in “risk” - the amount by which the value fluctuates. The most common measurement used is standard deviation and (its theoretically-nicer-but conceptually-wierder-brother) variance. We would like to be able to estimate a population standard deviation, to perform tests on a standard deviation, and to test for equality (or inequality) of standard deviations from two populations - but the work is carried out in terms of variance. Our best estimator for σ^2 is s^2 , if we look at all samples of size n, the mean of all the s^2 values will be exactly σ^2. [Because of the behavior of square roots, the values of s would not average out to the value of σ]. Distribution: The distribution of s^2 values is not normal — we can’t convert to a Z or t. If the variable (X) being observed is approximately normal, then the distribution of values of s^2 does fit the χ^2 (chi-square) family. We tend to get a χ^2 distribution if we add the squares of a lot of normal variables (as we do, in calculating s^2 , if X is normal). As with the t distributions, there is a different version of the χ^2 for each value of “‘degrees of freedom”; the distribution is not symmetric, and the mean gets larger with larger degrees of freedom. Table: We will work with a table of critical values for χ^2 — it’s a little more complicated than the t table because all values are positive (no ±tα tricks can be used) and the distribution is not symmetric. The value which cuts off area α on the right (larger) side is called χ^2 α (same as with Z or t) but the value that cuts off area α on the left can’t be −χ^2 α - we notice that it must cut off area 1 − α to the em right and call it χ^2 (1−α)

Test Statistic: To work with s^2 we use a normalization — but instead of subtracting the mean of the distribution, we divide by the mean (of the s^2 values — that is σ^2 of the X values) and multiply by degrees of freedom - so our test

statistic will be sample χ^2 = (n−1)s

2 σ^2.

Estimation We do not get an estimate of the form (statistic) ± E - we will be looking at multiplication/division: To estimate σ^2 , with confidence 1 − α we use the interval

(n − 1)s^2 χ^2 α/ 2

to (n − 1)s^2 χ^2 (1−α/2)

with df = n − 1

Tests For tests, we have the usual six steps, but the hypotheses are about σ^2 , the test statistic is sample χ^2 = (n−1)s

2 σ^2 and we have to look for χ^2 (1−α) or χ^2 (1−α/2) where we would have used −tα or −tα/ 2 for tests on a mean.

“Greater” H 0 : σ^2 = σ^20 Ha : σ^2 > σ 02

“Less” H 0 : σ^2 = σ^20 Ha : σ^2 < σ 02

“not equal” H 0 : σ^2 = σ^20 Ha : σ^2 6 = σ^20

sample χ^2 =

(n − 1)s^2 σ 02 df = n − 1

Reject H 0 if sample χ^2 > χ^2 α Reject H 0 if sample χ^2 < χ^2 (1−α) Reject H 0 if sample χ^2 < χ^2 (1− α 2 )

or sample χ^2 > χ^2 (1− α 2 )

Tests on difference of two variances To decide whether the variances of two populations are equal, we will consider the ratio of sample variances s

(^21) s^22.^ The distribution of this ratio will follow yet another distribution - this one for a ratio of χ^2 distributions. It’s called the F distribution and changes based on not one but two values - degrees of freedom for the numerator and degrees of freedom for the denominator. In this note I will use the notation F(1−α) ab for “the value of F , with degrees of freedom a for the numerator, v for the denominator, that cuts off area α to the right”. In your text (pp. 925-927) there is a table of critical values for the right-side only (that is Fα ab for small α). The conversion for the left-side value is not pretty but not terribly hard: The left-side critical value (for “ <′′^ and “ 6 =′′^ tests) is given by F(1−α) ab = (^) Fα^1 ba.

For the tests, we have the usual six steps, but the hypotheses are about σ 12 and σ^22 , the test statistic is sample F = s

(^21) s^22 and we have to look for F(1−α) ab or F(1−α/ 2 ) ab where we would have used −tα or −tα/ 2 for a test on means.

“Greater” H 0 : σ^21 = σ^22 Ha : σ 12 > σ 22

“Less” H 0 : σ^21 = σ^22 Ha : σ 12 < σ 22

“not equal” H 0 : σ^21 = σ^22 Ha : σ 12 6 = σ^22

sample F = s

(^21) s^22 numerator df = a = n 1 − 1 , denominator df = b = n 2 − 1

Reject H 0 if sample F > Fα ab Reject H 0 if sample F < F(1−α) ab Reject H 0 if sample F < F(1− α 2 ) ab or sample F > F(1− α 2 ) ab

Class examples for tests and estimation on variances.

  1. We want to estimate the risk involved in purchasing shares in Fly-By Night Inc. In particular we, would like the risk, as measured by standard deviation of the price, to be less than $10.00 (per share). A sample of prices for 22 trading days gives a mean price $31.50 and standard deviation $7.25, and it seems reasonable (based on past performance) to expect prices to be approximately normal.

(a) Does this give evidence that the standard deviation of prices is less than $10.00? (b) What is our 95% confidence estimate of the standard deviation of prices.

  1. A bottling machine is supposed to fill bottles with an average of 32.01 oz of soda, with standard deviation no more than .52 oz. A sample of 41 bottles of soda gives a mean fill 32.03 oz., standard deviation .62 oz.

(a) Does this indicate at the .01 level that the variation in amounts is greater than specifications? (b) What is our 95% confidence estimate for the variance of the amounts?

  1. We are comparing two types of thermostats — we want to know if either one provides a more even temperature than the other. We run a series of tests in which the thermostats are used to control the temperature of a room, which is supposed to be kept at 72 degrees Fahrenheit, and we take readings of the actual temperature at regular intervals. We will test for a difference in variability. For the “super climate” thermostat, we obtain 45 readings with mean 73.95 degrees and standard deviation 2.56 degrees. For the “heat master” we obtain 30 readings with mean 71.96 degrees and standard deviation 3.82 degrees. Does this give evidence (use .05 level) that there is a difference between the two thermostats in the variability of the temperatures?
  2. We want to know if investment in stock A carries less risk (lower standard deviation in price) than stock B. We have data for 31 days of prices of stock A, giving mean $38.67 and standard deviation $3.79 and for 26 days of prices of stock B, giving mean $42.75 and standard deviation $5.55. Does this data give evidence that the standard deviation of prices for A is less than the standard deviation for B?