on thursday, john will provide information about the project next homework: john will announce in...
TRANSCRIPT
• On Thursday, John will provide information about the project
• Next homework: John will announce in class on Thursday, and it’ll be due the following Thursday.
• Please hand in this week’s homework before you leave.
Announcements
Hypothesis Testing: 20,000 Foot View
1. Set up the hypothesis to test and collect data
Hypothesis to test: HO
Hypothesis Testing: 20,000 Foot View
1. Set up the hypothesis to testand collect data
2. Assuming that the hypothesis is true, are the observed data likely?
Data are deemed “unlikely” if the test statistics is in theextreme of its distribution when HO is true.
Hypothesis to test: HO
Hypothesis Testing: 20,000 Foot View
1. Set up the hypothesis to testand collect data
2. Assuming that the hypothesis is true, are the observed data likely?
3. If not, then the alternative to the hypothesis must be true.
Data are deemed “unlikely” if the test statistics is in theextreme of its distribution when HO is true.
Alternative to HO is HA
Hypothesis to test: HO
Hypothesis Testing: 20,000 Foot View
1. Set up the hypothesis to testand collect data
2. Assuming that the hypothesis is true, are the observed data likely?
3. If not, then the alternative to the hypothesis must be true.
4. P-value describes how likely the observed data are assuming HO is true. (i.e. answer to Q#2 above)
Data are deemed “unlikely” if the test statistics is in theextreme of its distribution when HO is true.
“Unlikely” if p-value < (smaller p-value = less likely)
Alternative to HO is HA
Hypothesis to test: HO
Large Sample Test for a Proportion:Taste Test Data
• 33 people drink two unlabeled cups of cola (1 is coke and 1 is pepsi)
• p = proportion who correctly identify drink
= 20/33 = 61%
• Question: is this statistically significantly different from 50% (random guessing) at = 10%?
Large Sample Test for a Proportion:Taste Test Data
• HO: p = 0.5HA: p does not equal 0.5
• Test statistic:z = | (p - .5)/sqrt( p(1-p)/n) |
= | (.61-.5)/sqrt(.61*.39/33) | = 1.25
• Reject if z > z0.10/2 = 1.645
• It’s not, so there’s not enough evidence to reject HO.
Large Sample Test for a Proportion:Taste Test Data
P-valuePr( |(P-p)/sqrt(P Q/n)| >
|(p-p)/sqrt(p q/n)| when H0 is true)=Pr( |(P-0.5)/sqrt(P Q/n) | > |1.25 | when H0 is true)=2*Pr( Z > 1.25) where Z~N(0,1)= 21%
i.e. “How likely is a test statistic of 1.25 when true p = 50%?”
The test
• The test statistic is:
z = | (p - .5)/sqrt( .5(1-.5)/n) |= | (.61-.5)/sqrt(.25/33) | = 1.22
Note that since .25 >= p(1-p) for any p, this is more conservative (larger denominator = smaller test statistic). Either way is fine.
Difference between two means
• PCB Data– Sample 1: Treatment is to expose cells to a certain
PCB (number 156)– Sample 2: Treatment is to expose the cells to PCB
156 and another chemical compound (estradiol)
• Response = estrogen produced by cells• Question: Can we conclude that average
estrogen produced in sample 1 is different from average by sample 2 (at = 0.05)?
• H0: 1 – 2 = 0HA: 1 – 2 does not = 0
• Test statistic:|(Estimate – value under H0)/Std Dev(Estimate)|
z = (x1 – x2)/sqrt(s12/n1 + s2
2/n2)
Reject if |z| > z/2
• P-value = 2*Pr[ Z > (x1 – x2)/sqrt(s1
2/n1 + s22/n2)] where
Z~N(0,1).
Form of the test
n x s
PCB156 96 1.93 1.00
PCB156+E 64 2.16 1.01
|z| = |-0.229/sqrt(1.002/96 + 1.012/64)| = |-1.41| = 1.41
z/2 = z0.05/2 = z0.025 = 1.96
So don’t reject.
P-value = 2*Pr(Z > 1.41) = 16%
Pr( Test statistic > 1.41 when HO is true)
Data
• Test statistic: |z| = |(Estimate – assumed value under H0)/(Std Dev of the Estimator)| (note that Std Dev of the estimator is the Standard Dev of the individual data points that go into the estimators divided by the square root of n)
• Reject if |z| > z/2
• P-value = 2*Pr( Z > z ) where Z~N(0,1).In the previous example, what was the estimator? What was its
standard error? Note the similarity to a confidence interval for the difference between two means.
In General, Large Sample 2 sided Tests:
Large Sample Hypothesis Tests: summary for means
Single meanHypotheses Test (level 0.05)
HO: = k Reject HO if |(x-k)/s/sqrt(n)|>1.96
HA: does not = k p-value: 2*Pr(Z>|(x-k)/s/sqrt(n)|)where Z~N(0,1)
Difference between two meansHypotheses Test (level 0.05)
HO: = D Let d = x1 – x2
HA: does not = D Let SE = sqrt(s12/n2 + s2
2/n2)Reject HO if |(d-D)/SE|>1.96
p-value: 2*Pr(Z>|(d-D)/SE|)where Z~N(0,1)
Large Sample Hypothesis Tests: summary for proportions
Single proportionHypotheses Test (level 0.05)
HO: true p = k Reject HO if |(p-k)/sqrt(p(1-p)/n)|>1.96
HA: p does not = k p-value: 2*Pr(Z>|(p-k)/sqrt(p(1-p)/n)|)where Z~N(0,1)
Difference between two proportionsHypotheses Test (level 0.05)
HO: p1-p2 = 0 Let d = p1 – p2
HA: p1-p2 does not = 0 Let p = total “success”/(n1+n2)Let SE = sqrt(p(1-p)/n1 + p(1-p)/n2)
Reject HO if |(d)/SE|>1.96p-value: 2*Pr(Z>|(d)/SE|)
where Z~N(0,1)
• A two sided level hypothesis test, H0: =k vs HA: does not equal k
is rejected if and only if k is not in a 1- confidence interval for the mean.
• A one sided level hypothesis test, H0: <=k vs HA: >k
is rejected if and only if a level 1-2 confidence interval is completely to the left of k.
Hypothesis tests versus confidence intervals
The following is discussed in the context of tests / CI’s for a single mean, but it’s true for all the confidence intervals / tests we have done.
• The previous slide said that confidence intervals can be used to do hypothesis tests.
• CI’s are “better” since they contain more information.
• Fact: Hypothesis tests and p-values are very commonly used by scientists who use statistics.
• Advice: 1. Use confidence intervals to do hypothesis testing2. know how to compute / and interpret p-values
Hypothesis tests versus confidence intervals
Project and Exam 2:• Work in groups. (preferably 2 or 3 people)• Collect and analyze data.
– Analysis should include: summary statistics and good graphical displays.
– Also, should include at least one of the following: confidence intervals, hypothesis tests, analysis of variance, regression, contingency tables. In many cases power calculations will be required too (I’ll tell you after reading your proposals).
– Write-up: General description of what you are doing and questions the statistics will address. Statistics and interpretation. (<6 pages)
• One page proposal (who and what) due this Thursday. Project due May 13th.
• Exam 2 is optional. If you want to take it, contact me to schedule a time.
We talked about Type 1 and Type 2 Errors generally before, now we’ll compute some of
the associated probabilities.
Truth
H0 True
HA True
Action
Fail to Reject H0 Reject H0
correct
correct
Type 1error
Type 2error
Significance level = =Pr(Making type 1 error)
Power = 1–Pr(Making type 2 error)
Example: Dietary Folate• Data from the Framingham Heart Study
0 200 400 600 800 1000 1200
0
20
40
60
80
100
Dietary Folate (micrograms / day, calorie adjusted to 2000 calorie diet)
Count
n = 333 Elderly Men
Mean = x = 336.4
Std Dev = s = 193.4
Can we conclude that the mean is greater than 300 at 5% significance? (same as 95% confidence)
In terms of our folate example, suppose we repeated the experiment
and sampled 333 new people
Pr( Type 1 error )= Pr( reject H0 when mean is 300 )= Pr( |Z| > z0.025 )= Pr( Z > 1.96 ) + Pr( Z < -1.96 ) = 0.05 =
When mean is 300, then Z, the test statistic, has a standard normal distribution.
Note that the test is designed to have type 1 error =
Power= Pr( reject H0 when mean is not 300 )= Pr( reject H0 when mean is 310)= Pr( |(X-300)/193.4/sqrt(333)| > 1.96)= Pr( (X-300)/10.6 > 1.96 )+Pr( (X-300)/10.6 < -1.96 )= Pr(X > 320.8) + Pr(X < 279.2)
= Pr( (X – 310)/10.6 > (320.8-310)/10.6 )+ Pr( (X – 310)/10.6 < (279.2-310)/10.6 )
= Pr( Z > 1.02 ) + Pr( Z < -2.90 ) where Z~N(0,1)= 0.15 + 0.00 = 0.15
What’s the power to detect an increase ofof least 10 units of folate for a new experiment? (using a 2 sided test for a single mean)
In other words, if the true mean is 310 and the standard error is 10.6, then there’s an 85% chance that we will not detect it in a new experiment at the 5% level.
If 310 is scientifically significantly different from 300, then this means that our experiment is likely to be wasted.
If all else is held constant, then:
As n increases, power goes up.
As standard deviation of x decreases, power goes up.
As increases, power goes up.
Picture for Power
True Mean
Po
we
r
260 280 300 320 340
0.2
0.4
0.6
0.8
1.0
Power forn=333 and = 0.05
“Pr(Reject HO when it’s false)”
As n increases and/or increases and/or stddev decreases, these
curves becomesteeper
Power calculations are an integral part of planning any experiment:
• Given:– a certain level of – preliminary estimate of std dev (of x’s that go
into x)– difference that is of interest
• Compute required n in order for power to be at least 85% (or some other percentage...)
“Post Hoc Power”
• Note that it is nonsense to do a power calculation after you have collected the data and conclude that you “would’ve seen significant results if you’d had a larger sample size (or lower standard deviation, etc).
• The fallacy is you would get the same x-bar if you collected more data. After the experiment, you only know that x-bar is likely to be in its confidence interval. You do not know where!
Power calculations are an integral part of planning any experiment:
• Bad News: Algebraically messy (but you should know how to do them)
• Good News: Minitab can be used to do them:• Menu: Stat: Power and Sample Size…
– Inputs:1. required power
2. difference of interest
– Output:Result = required sample size
– Options: Change , one sided versus 2 sided tests
Inference from Small SamplesChapter 10
• Data from a manufacturer of child’s pajamas
• Want to develop materials that take longer before they burn.
• Run an experiment to compare four types of fabrics. (They considered other factors too, but we’ll only consider the fabrics. Source: Matt Wand)
4321
18
17
16
15
14
13
12
11
10
9
Fabric
Bur
n T
ime
Fabric Data:Tried to light 4 samples of 4 different (unoccupied!) pajama fabrics on fire.
Higher #meanslessflamable
Mean=16.85std dev=0.94
Mean=10.95std dev=1.237 Mean=10.50
std dev=1.137
Mean=11.00std dev=1.299
Confidence Intervals?
• Suppose we want to make confidence intervals of mean “burn time” for each fabric type.
• Can I use: x +/- z/2s/sqrt(n) for each one?
• Why or why not?
Answer:
tn-1 is the “t distribution” with n-1 degrees of freedom (df)
• Sample size (n=4) is too small to justify central limit theorem based normal approximation.
• More precisely:– If xi is normal, then (x – )/[/sqrt(n)] is normal for any n.
– xi is normal, then (x – )/[s/sqrt(n)] is normal for n > 30.
– New: Suppose xi is approximately normal (and an independent sample). Then (x – )/[s/sqrt(n)] ~ tn-1
Parameter: (number of data points used to estimate s) - 1
“Student” t-distribution(like a normal distribution, but w/ “heavier tails”)
x
dens
ity
-4 -2 0 2 4
0.0
0.1
0.2
0.3
0.4
t dist’t with 3df
Normal dist’n
As df increases, tn-1 becomes the normal dist’n. Indistinguishable for n > 30 or so.
Idea: estimating stddev leads to “morevariability”. Morevariability = higher chance of “extreme”observation
t-based confidence intervals
• 1- level confidence interval for a mean:
x +/- t/2,n-1s/sqrt(n)
where t/2,n-1 is a number such thatPr(T > t/2,n-1) = /2 and T~tn-1
(see table opposite normal table inside of book cover…)
Back to burn time example
x s t0.025,3 95% CI
Fabric 1 16.85 0.940 3.182 (15.35,18.35)
Fabric 2 10.95 1.237 3.182 (8.98, 12.91)
Fabric 3 10.50 1.137 3.182 (8.69, 12.31)
Fabric 4 11.00 1.299 3.182 (8.93, 13.07)
t-based Hypothesis test for a single mean
• Mechanics: replace z/2 cutoff with t/2,n-1ex: fabric 1 burn time dataH0: mean is 15HA: mean isn’t 15Test stat: |(16.85-15)/(0.94/sqrt(4))| = 3.94Reject at =5% since 3.94>t0.025,3=3.182P-value = 2*Pr(T>3.94) where T~t3. This is between 2% and 5% since t0.025,3=3.182 and t0.01,3=4.541. (pvalue=2*0.0146) from software)
• See minitab: basis statistics: 1 sample t test• Idea: t-based tests are harder to pass than large
sample normal based test. Why does that make sense?