Statistical Inference I: Hypothesis testing; sample size

Preview:

Citation preview

Statistical Inference I: Hypothesis testing; sample

size

Statistics Primer Statistical Inference Hypothesis testing P-values Type I error Type II error Statistical power Sample size calculations

What is a statistic? A statistic is any value that can

be calculated from the sample data.

Sample statistics are calculated to give us an idea about the larger population.

Examples of statistics: mean

The average cost of a gallon of gas in the US is $2.65. difference in means

The difference in the average gas price in Los Angeles ($2.96) compared with Des Moines, Iowa ($2.32) is 64 cents.

proportion 67% of high school students in the U.S. exercise

regularly difference in proportions

The difference in the proportion of Democrats who approve of Obama (83%) versus Republicans who do (14%) is 69%

What is a statistic? Sample statistics are estimates of

population parameters.

Sample (observation)

Make guesses about the whole population

Sample statistics estimate population parameters:

Truth (not observable)

Mean IQ of some population

of 100,000 people =100

Sample statistic: mean IQ of 5 subjects

1105

11512496105110

What is sampling variation?

Statistics vary from sample to sample due to random chance.

Example: A population of 100,000 people has an

average IQ of 100 (If you actually could measure them all!)

If you sample 5 random people from this population, what will you get?

Truth (not observable)

1305

9595180160120

Sampling Variation

Mean IQ=100

905

8892958590

1105

11512496105110

985

9510486105100

Sampling Variation and Sample Size Do you expect more or less

sampling variability in samples of 10 people?

Of 50 people? Of 1000 people? Of 100,000 people?

Most experiments are one-shot deals. So, how do we know if an observed effect from a single experiment is real or is just an artifact of sampling variability (chance variation)?

 **Requires a priori knowledge about how sampling variability

works…

Question: Why have I made you learn about probability distributions and about how to calculate and manipulate expected value and variance?

Answer: Because they form the basis of describing the distribution of a sample statistic.

Sampling Distributions

Standard error Standard Error is a measure of sampling

variability. Standard error is the standard deviation of a

sample statistic. It’s a theoretical quantity! What would the distribution of my statistic be if I

could repeat my experiment many times (with fixed sample size)? How much chance variation is there?

Standard error decreases with increasing sample size and increases with increasing variability of the outcome (e.g., IQ).

Standard errors can be predicted by computer simulation or mathematical theory (formulas).

The formula for standard error is different for every type of statistic (e.g., mean, difference in means, odds ratio).

What is statistical inference? The field of statistics provides

guidance on how to make conclusions in the face of chance variation (sampling variability).

Example 1: Difference in proportions Research Question: Are

antidepressants a risk factor for suicide attempts in children and adolescents?

Example modified from: “Antidepressant Drug Therapy and Suicide in Severely Depressed Children and Adults ”; Olfson et al. Arch Gen Psychiatry.2006;63:865-872.

Example 1 Design: Case-control study Methods: Researchers used Medicaid

records to compare prescription histories between 263 children and teenagers (6-18 years) who had attempted suicide and 1241 controls who had never attempted suicide (all subjects suffered from depression).

Statistical question: Is a history of use of antidepressants more common among cases than controls?

Example 1 Statistical question: Is a history of use of

particular antidepressants more common among heart disease cases than controls?

What will we actually compare? Proportion of cases who used

antidepressants in the past vs. proportion of controls who did

No (%) of cases

(n=263)

No (%) of controls (n=1241)

Any antidepressant drug ever 120 (46%)  448 (36%)

46% 36%

Difference=10%

Results

What does a 10% difference mean?

Before we perform any formal statistical analysis on these data, we already have a lot of information.

Look at the basic numbers first; THEN consider statistical significance as a secondary guide.

Is the association statistically significant? This 10% difference could reflect a

true association or it could be a fluke in this particular sample.

The question: is 10% bigger or smaller than the expected sampling variability?

What is hypothesis testing? Statisticians try to answer this

question with a formal hypothesis test

Hypothesis testing

Null hypothesis: There is no association between antidepressant use and suicide attempts in the target population (= the difference is 0%)

Step 1: Assume the null hypothesis.

Hypothesis Testing

Step 2: Predict the sampling variability assuming the null hypothesis is true—math theory (formula):

The standard error of the difference in two proportions is:

Thus, we expect to see differences between the group as big as about 6.6% (2 standard errors) just by chance…

033.1241

)1504568

1(1504568

263

)1504568

1(1504568

)-1()1(

21

n

pp

n

pp

Hypothesis Testing

In computer simulation, you simulate taking repeated samples of the same size from the same population and observe the sampling variability.

I used computer simulation to take 1000 samples of 263 cases and 1241 controls assuming the null hypothesis is true (e.g., no difference in antidepressant use between the groups).

Step 2: Predict the sampling variability assuming the null hypothesis is true—computer simulation:

Computer Simulation Results

What is standard error?

Standard error: measure of variability of sample statistics

Standard error is about 3.3%

Hypothesis Testing

Step 3: Do an experiment

We observed a difference of 10% between cases and controls.

Hypothesis Testing

Step 4: Calculate a p-value

P-value=the probability of your data or something more extreme under the null hypothesis.

Hypothesis Testing

Step 4: Calculate a p-value—mathematical theory:

003.=p;0.3=033.

10.=Z

Observed difference between the groups.

Standard error.

Difference in proportions follows a normal distribution.

A Z-value of 3.0 corresponds to a p-value of .003.

When we ran this study 1000 times, we got 1 result as big or bigger than 10%.

The p-value from computer simulation…

We also got 2 results as small or smaller than –10%.

P-valueP-value

P-value=the probability of your data or something more extreme under the null hypothesis.

From our simulation, we estimate the p-value to be:

3/1000 or .003

Here we reject the null.

Alternative hypothesis: There is an association between antidepressant use and suicide in the target population.

Hypothesis Testing

Step 5: Reject or do not reject the null hypothesis.

What does a 10% difference mean? Is it “statistically significant”? YES Is it clinically significant? Is this a causal association?

What does a 10% difference mean? Is it “statistically significant”? YES Is it clinically significant? MAYBE Is this a causal association? MAYBE

Statistical significance does not necessarily imply clinical significance.

Statistical significance does not necessarily imply a cause-and-effect relationship.

What would a lack of statistical significance mean?

If this study had sampled only 50 cases and 50 controls, the sampling variability would have been much higher—as shown in this computer simulation…

Standard error is about 10%

50 cases and 50 controls.

Standard error is about 3.3% 263 cases and

1241 controls.

With only 50 cases and 50 controls…

Standard error is about 10%

If we ran this study 1000 times, we would expect to get values of 10% or higher 170 times (or 17% of the time).

Two-tailed p-value

Two-tailed p-value = 17%x2=34%

What does a 10% difference mean (50 cases/50 controls)? Is it “statistically significant”? NO Is it clinically significant? MAYBE Is this a causal association? MAYBE

No evidence of an effect Evidence of no effect.

Example 2: Difference in means

Example: Rosental, R. and Jacobson, L. (1966) Teachers’ expectancies: Determinates of pupils’ I.Q. gains. Psychological Reports, 19, 115-118.

The Experiment (note: exact numbers have been altered)

Grade 3 at Oak School were given an IQ test at the beginning of the academic year (n=90).

Classroom teachers were given a list of names of students in their classes who had supposedly scored in the top 20 percent; these students were identified as “academic bloomers” (n=18).

BUT: the children on the teachers lists had actually been randomly assigned to the list.

At the end of the year, the same I.Q. test was re-administered.

Example 2 Statistical question: Do students in the

treatment group have more improvement in IQ than students in the control group?

What will we actually compare? One-year change in IQ score in the

treatment group vs. one-year change in IQ score in the control group.

“Academic bloomers”

(n=18)Controls (n=72)

Change in IQ score: 12.2 (2.0)  8.2 (2.0)

Results:

12.2 points 8.2 points

Difference=4 points

The standard deviation of change scores was 2.0 in both groups. This affects statistical significance…

What does a 4-point difference mean?

Before we perform any formal statistical analysis on these data, we already have a lot of information.

Look at the basic numbers first; THEN consider statistical significance as a secondary guide.

Is the association statistically significant? This 4-point difference could reflect

a true effect or it could be a fluke. The question: is a 4-point

difference bigger or smaller than the expected sampling variability?

Hypothesis testing

Null hypothesis: There is no difference between “academic bloomers” and normal students (= the difference is 0%)

Step 1: Assume the null hypothesis.

Hypothesis Testing

Step 2: Predict the sampling variability assuming the null hypothesis is true—math theory:

52.072

4

18

4

2

2

1

2

n

s

n

s

The standard error of the difference in two means is:

We expect to see differences between the group as big as about 1.0 (2 standard errors) just by chance…

Hypothesis Testing

In computer simulation, you simulate taking repeated samples of the same size from the same population and observe the sampling variability.

I used computer simulation to take 1000 samples of 18 treated and 72 controls, assuming the null hypothesis (that the treatment doesn’t affect IQ).

Step 2: Predict the sampling variability assuming the null hypothesis is true—computer simulation:

Computer Simulation Results

What is the standard error?

Standard error: measure of variability of sample statistics

Standard error is about 0.52

Hypothesis Testing

Step 3: Do an experiment

We observed a difference of 4 between treated and controls.

Hypothesis Testing

Step 4: Calculate a p-value

P-value=the probability of your data or something more extreme under the null hypothesis.

Hypothesis Testing

 

p-value <.0001852.

488 t

Step 4: Calculate a p-value—mathematical theory:

Difference in means follows a T distribution (which is very similar to a normal except with very small samples).

Observed difference between the groups.

Standard error.

A T88-value of 8.0 corresponds to a p-value of <.0001

A t-curve with 88 df’s has slightly wider cut-off’s for 95% area (t=1.99) than a normal curve (Z=1.96)

If we ran this study 1000 times, we wouldn’t expect to get 1 result as big as a difference of 4 (under the null hypothesis).

Getting the P-value from computer simulation…

P-value=the probability of your data or something more extreme under the null hypothesis.

Here, p-value<.0001

P-valueP-value

Here we reject the null.

Alternative hypothesis: There is an association between being labeled as gifted and subsequent academic achievement.

Hypothesis Testing

Step 5: Reject or do not reject the null hypothesis.

What does a 4-point difference mean? Is it “statistically significant”? YES Is it clinically significant? Is this a causal association?

What does a 4-point difference mean? Is it “statistically significant”? YES Is it clinically significant? MAYBE Is this a causal association? MAYBE

Statistical significance does not necessarily imply clinical significance.

Statistical significance does not necessarily imply a cause-and-effect relationship.

What if our standard deviation had been higher? The standard deviation for change

scores in both treatment and control was 2.0. What if change scores had been much more variable—say a standard deviation of 10.0?

Standard error is 0.52 Std. dev in

change scores = 2.0

Std. dev in change scores = 10.0

Standard error is 2.58

With a std. dev. of 10.0…

Standard error is 2.58

If we ran this study 1000 times, we would expect to get +4.0 or –4.0 12% of the time.

P-value=.12

What would a 4.0 difference mean (std. dev=10)? Is it “statistically significant”? NO Is it clinically significant? MAYBE Is this a causal association? MAYBE

No evidence of an effect Evidence of no effect.

Hypothesis testing summary Null hypothesis: the hypothesis of no effect

(usually the opposite of what you hope to prove). The straw man you are trying to shoot down.

Example: antidepressants have no effect on suicide risk P-value: the probability of your observed data if

the null hypothesis is true. Example: The probability that the study would have found

10% higher suicide attempts in the antidepressant group (compared with control) if antidepressants had no effect (i.e., just by chance).

If the p-value is low enough (i.e., if our data are very unlikely given the null hypothesis), this is evidence that the null hypothesis is wrong.

If p-value is low enough (typically <.05), we reject the null hypothesis and conclude that antidepressants do have an effect.

Summary: The Underlying Logic of hypothesis tests…

Follows this logic:

Assume A.

If A, then B.

Not B.

Therefore, Not A.

But throw in a bit of uncertainty…If A, then probably B…

Error and power Type I error rate (or significance level): the

probability of finding an effect that isn’t real (false positive).

If we require p-value<.05 for statistical significance, this means that 1/20 times we will find a positive result just by chance.

Type II error rate: the probability of missing an effect (false negative).

Statistical power: the probability of finding an effect if it is there (the probability of not making a type II error).

When we design studies, we typically aim for a power of 80% (allowing a false negative rate, or type II error rate, of 20%).

Type I and Type II Error in a box

Your Statistical Decision

True state of null hypothesis

H0 True (e.g., antidepressants do not increase suicide risk)

H0 False(e.g., antidepressants do

increase suicide risk)

Reject H0

(e.g., conclude antidepressants increase suicide risk)

Type I error (α) Correct

Do not reject H0(e.g., conclude there is insufficient evidence that antidepresssants increase suicide risk)

Correct Type II Error (β)

Reminds me of…

Pascal’s Wager

Your DecisionThe TRUTH

God Exists God Doesn’t Exist

Reject GodBIG MISTAKE Correct

Accept God Correct—Big Pay Off

MINOR MISTAKE

Type I and Type II Error in a box

Your Statistical Decision

True state of null hypothesis

H0 True H0 False

Reject H0 Type I error (α) Correct

Do not reject H0

Correct Type II Error (β)

Review Question 1If we have a p-value of 0.03 and so decide that our effect is statistically significant, what is the probability that we’re wrong (i.e., that the hypothesis test gave us a false positive)?

a. .03b. .06c. Cannot telld. 1.96e. 95%

Review Question 1If we have a p-value of 0.03 and so decide that our effect is statistically significant, what is the probability that we’re wrong (i.e., that the hypothesis test gave us a false positive)?

a. .03b. .06c. Cannot telld. 1.96e. 95%

Review Question 2Standard error is:

a. For a given variable, its standard deviation divided by the square root of n.

b. A measure of the variability of a sample statistic.

c. The inverse of sample size. d. A measure of the variability of a

characteristic.e. All of the above.

Review Question 2Standard error is:

a. For a given variable, its standard deviation divided by the square root of n.

b. A measure of the variability of a sample statistic.

c. The inverse of sample size. d. A measure of the variability of a

characteristic.e. All of the above.

Review Question 3A randomized trial of two treatments for depression failed to show a statistically significant difference in improvement from depressive symptoms (p-value =.50). It follows that:

a. The treatments are equally effective.b. Neither treatment is effective.c. The study lacked sufficient power to detect a

difference.d. The null hypothesis should be rejected.e. There is not enough evidence to reject the null

hypothesis.

Review Question 3A randomized trial of two treatments for depression failed to show a statistically significant difference in improvement from depressive symptoms (p-value =.50). It follows that:

a. The treatments are equally effective.b. Neither treatment is effective.c. The study lacked sufficient power to detect a

difference.d. The null hypothesis should be rejected.e. There is not enough evidence to reject the null

hypothesis.

Review Question 4Following the introduction of a new treatment regime in a rehab facility, alcoholism “cure” rates increased. The proportion of successful outcomes in the two years following the change was significantly higher than in the preceding two years (p-value: <.005).  It follows that:

a. The improvement in treatment outcome is clinically important.b. The new regime cannot be worse than the old treatment.c. Assuming that there are no biases in the study method, the new

treatment should be recommended in preference to the old. d. All of the above.e. None of the above.

Review Question 4Following the introduction of a new treatment regime in a rehab facility, alcoholism “cure” rates increased. The proportion of successful outcomes in the two years following the change was significantly higher than in the preceding two years (p-value: <.005).  It follows that:

a. The improvement in treatment outcome is clinically important.b. The new regime cannot be worse than the old treatment.c. Assuming that there are no biases in the study method, the new

treatment should be recommended in preference to the old. d. All of the above.e. None of the above.

Statistical Power Statistical power is the probability

of finding an effect if it’s real.

Can we quantify how much power we have for given sample sizes?

Null Distribution: difference=0.

Clinically relevant alternative: difference=10%.

Rejection region. Any value >= 6.5 (0+3.3*1.96)

study 1: 263 cases, 1241 controls

For 5% significance level, one-tail area=2.5%

(Z/2 = 1.96)

Power= chance of being in the rejection region if the alternative is true=area to the right of this line (in yellow)

Power here = >80%

Rejection region. Any value >= 6.5 (0+3.3*1.96)

Power= chance of being in the rejection region if the alternative is true=area to the right of this line (in yellow)

study 1: 263 cases, 1241 controls

Critical value= 0+10*1.96=20

Power closer to 20% now.

2.5% area

Z/2=1.96

study 1: 50 cases, 50 controls

Critical value= 0+0.52*1.96 = 1

Power is nearly 100%!

Study 2: 18 treated, 72 controls, STD DEV = 2

Clinically relevant alternative: difference=4 points

Critical value= 0+2.59*1.96 = 5

Power is about 40%

Study 2: 18 treated, 72 controls, STD DEV=10

Critical value= 0+0.52*1.96 = 1

Power is about 50%

Study 2: 18 treated, 72 controls, effect size=1.0

Clinically relevant alternative: difference=1 point

Factors Affecting Power

1. Size of the effect2. Standard deviation of the

characteristic3. Bigger sample size 4. Significance level desired

average weight from samples of 100

Null

Clinically relevant alternative

1. Bigger difference from the null mean

average weight from samples of 100

2. Bigger standard deviation

average weight from samples of 100

3. Bigger Sample Size

average weight from samples of 100

4. Higher significance level

Rejection region.

Sample size calculations Based on these elements, you can

write a formal mathematical equation that relates power, sample size, effect size, standard deviation, and significance level…

Simple formula for difference in proportions

221

2/2

)(p

)Z)(1)((2

p

Zppn

Sample size in each group (assumes equal sized groups)

Represents the desired power (typically .84 for 80% power).

Represents the desired level of statistical significance (typically 1.96).

A measure of variability (similar to standard deviation)

Effect Size (the difference in proportions)

Simple formula for difference in means

Sample size in each group (assumes equal sized groups)

Represents the desired power (typically .84 for 80% power).

Represents the desired level of statistical significance (typically 1.96).

Standard deviation of the outcome variable

Effect Size (the difference in means)

2

2/2

2

difference

)Z(2

Zn

Sample size calculators on the web… http://biostat.mc.vanderbilt.edu/twi

ki/bin/view/Main/PowerSampleSize

http://calculators.stat.ucla.edu http://

hedwig.mgh.harvard.edu/sample_size/size.html

These sample size calculations are idealized

•They do not account for losses-to-follow up (prospective studies)

•They do not account for non-compliance (for intervention trial or RCT)

•They assume that individuals are independent observations (not true in clustered designs)

•Consult a statistician!

Review Question 5Which of the following elements does not

increase statistical power?

a. Increased sample sizeb. Measuring the outcome variable more

preciselyc. A significance level of .01 rather

than .05d. A larger effect size.

Review Question 5Which of the following elements does not

increase statistical power?

a. Increased sample sizeb. Measuring the outcome variable more

preciselyc. A significance level of .01 rather

than .05d. A larger effect size.

Review Question 6Most sample size calculators ask you to input a value for . What are they asking for?

a. The standard errorb. The standard deviationc. The standard error of the differenced. The coefficient of deviatione. The variance

Review Question 6Most sample size calculators ask you to input a value for . What are they asking for?

a. The standard errorb. The standard deviationc. The standard error of the differenced. The coefficient of deviatione. The variance

Review Question 7For your RCT, you want 80% power to detect

a reduction of 10 points or more in the treatment group relative to placebo. What is 10 in your sample size formula?

a. Standard deviationb. mean changec. Effect sized. Standard errore. Significance level

Review Question 7For your RCT, you want 80% power to detect

a reduction of 10 points or more in the treatment group relative to placebo. What is 10 in your sample size formula?

a. Standard deviationb. mean changec. Effect sized. Standard errore. Significance level

Homework Problem Set 3 Reading: continue reading

textbook Reading: p-value article Journal article/article review sheet

Recommended