Chapter 10: The t Test For Two Independent Samples

Chapter 10:
t Test for Two Independent Samples
Dr. John J. Kerbs, Associate Professor
Joint Ph.D. in Social Work and Sociology
Dr. John J. Kerbs can be emailed for permission at [email protected]
Independent-Measures Designs
Allows researchers to evaluate the mean difference
between two populations using data from two
separate samples.
 The identifying characteristic of the independentmeasures or between-subjects design is the
existence of two separate or independent
 Thus, an independent-measures design can be
used to test for mean differences between two
distinct populations (such as men versus women)
or between two different treatment conditions
(such as drug versus no-drug).
Teaching Method A vs. B
Independent-Measures Designs
The independent-measures design is used in
situations where a researcher has no prior
knowledge about either of the two populations
(or treatments) being compared.
In particular, the population means and standard
deviations are all unknown.
 Because the population variances are not known,
these values must be estimated from the sample
The t Statistic for an IndependentMeasures Research Design
As with all hypothesis tests, the general purpose of the
independent-measures t test is to determine whether the
sample mean difference obtained in a research study indicates
a real mean difference between the two populations (or
treatments) or whether the obtained difference is simply the
result of sampling error.
Remember, if two samples are taken from the same
population and are given exactly the same treatment, there will
still be some difference between the sample means (i.e., this
difference is called sampling error).
The hypothesis test provides a standardized, formal procedure
for determining whether the mean difference obtained in a
research study is significantly greater than can be explained
by sampling error
Two Population Distributions
Hypothesis Tests and Effect Size with
the Independent Measures t Statistic
To prepare the data for analysis, the first step
is to compute the sample mean and SS (or s,
or s2) for each of the two samples.
 The hypothesis test follows the same four-step
procedure outlined in Chapters 8 and 9.
Elements of a t statistics:
Single-Sample &
Independent Measures
NOTE: This alternative formula on the
bottom left for pooled variance works
when you have sample variances for
the first and second samples (s2) and
not sum of squares (SS) as noted here
Hypothesis Testing with the
Independent-Measures t Statistic
Book Example:
For 10 students who watched Sesame Street
(group 1) and 10 students who did not watch the
show (group 2), was there a difference in their
average high-school grades.
 Key Information to Run t-tests for independentmeasures
 n1
= 10
n2 = 10
 M1 = 93
M2 = 85
 SS1 = 200 SS2 = 160
Hypothesis Testing with the
Independent-Measures t Statistic
Step 1:
State the hypotheses and select the α level. For the
independent-measures test, H0 states that there is no
difference between the two population means.
For a two-tailed test, the hypotheses are as follows:
Example: H0: µ1 - µ2 = 0
H1: µ1 - µ2 ≠ 0
Select α: α = 0.01 (two tail)
For a one-tail test, H0: µ1 - µ2 ≤ 0
H1: µ1 - µ2 > 0
Hypothesis Testing with the
Independent-Measures t Statistic
Step 2:
Locate the critical region. The critical values for the t
statistic are obtained using degrees of freedom that are
determined by adding together the df value for the first
sample and the df value for the second sample. For two
samples with n = 10 in each sample, df is calculated as
df = df1 + df2
df = (n1 - 1) + (n2 -1)
df = 9 + 9 = 18
To find the critical t value for 18df at α = 0.01, we look at the
t-distribution table: t = +/- 2.878
Hypothesis Testing with the
Independent-Measures t Statistic
Step 3:
Compute the test statistic. The t statistic for the
independent-measures design has the same structure as
the single sample t introduced in Chapter 9. However, in
the independent-measures situation, all components of the t
formula are doubled: there are two sample means, two
population means, and two sources of error contributing to
the standard error in the denominator. Three key parts to
the t-statistic calculation:
Part A) Calculate the pooled variance
Part B) Use the pooled variance to calculate the estimated standard
Part C) Compute t-statistic
Hypothesis Testing with the
Independent-Measures t Statistic
Step 3 (Part A): Pooled Variance
Step 3 (Part B): Estimated Standard Error
Step 3 (Part C): Compute t-Statistic
1 − 2 −(µ1 − µ2 )
(1 − 2 )
1 − 2 −(µ1 − µ2 )
(1 − 2 )
93−85 −0
= =4
Hypothesis Testing with the
Independent-Measures t Statistic
Step 4:
Make a decision. If the t statistic ratio indicates that the
obtained difference between sample means (numerator) is
substantially greater than the difference expected by
chance (denominator), we reject H0 and conclude that there
is a real mean difference between the two populations or
T = 4.00>+2.878, thus reject H0 and conclude that there is
a significant difference in high school grades for those who
watched Sesame Street as compared to those who did not
watch this show.
Measuring Effect Size for the
Independent-Measures t
Effect size for the independent-measures t is
measured in the same way that we measured
effect size for the single-sample t in Chapter 9.
 Specifically, you can compute an estimate of
Cohen’s d or you can compute r2 to obtain a
measure of the percentage of variance
accounted for by the treatment effect.
Measuring Effect Size for the
Independent-Measures t
Compute an Estimated Cohen’s d
Est Cohen’s d =
Magnitude of d
1 − 2
= 1.79
Evaluation of Effect Size
d = 0.2
Small effect (mean difference around 0.2 standard deviations)
d = 0.5
Medium effect (mean difference around 0.5 standard deviations)
d = 0.8
Large effect (mean difference around 0.8 standard deviations)

 Compute r = 2
= 2
= 16
= 0.47 or 47%
4 +18
Percent of Variance Explained
as Measured
by r2 16+18
Evaluation of Effect Size
r2 = 0.01 (0.01*100 = 1%)
Small effect
r2 = 0.09 (0.09*100 = 9%)
Medium effect
r2 = 0.25 (0.25*100 = 25%)
Large effect
Removing Treatment Effects
NOTE: We added 4
points to those
students who did not
watch Sesame Street
subtracted 4 points
to those students
who watched
Sesame Street
Confidence Intervals and
Hypothesis Tests
Calculate a 95% Confidence Interval for the Sesame Street
example. For a two-tail critical value with df=18 and α = 0.05,
tcrit = +/- 2.101.
NOTE: Because 0 is not in the 95% C Interval,
we can conclude that the value of 0 is not in
µ1 - µ2 = M – M +/-t(1 − 2 )
the 95% Confidence Interval. Alternatively, the
= 93 – 85 +/- 2.101(2)
= 8 +/- 4.202
value of 0 is rejected with 95% confidence.
This is the same as rejecting H0 with α = 0.05
Sample Variance and Sample Size
Standard error is positively related to sample variance (larger
variance leads to larger sample error).
Standard error is inversely related to sample size (large
sample sizes lead to smaller standard error).
Larger variance produces smaller t-statistic and reduces the
likelihood of a significant finding.
 Larger variance also produces smaller measures of effect
Larger samples produces larger values for t-statistic and
increases the likelihood of rejecting H0.
 Sample size has no effect on Cohen’s d and only a small
influence on r2
Homogeneity of Variance Assumption
Most hypothesis tests usually work reasonably well
even if the set of underlying assumptions are violated.
The one notable exception is the assumption of
homogeneity of variance for the independentmeasures t test.
 Requires that the two populations from which the
samples are obtained have equal variances
 Necessary in order to justify pooling the two sample
variances and using the pooled variance in the
calculation of the t statistic
Homogeneity of Variance Assumption
If the assumption is violated, then the t statistic
contains two questionable values: (1) the value for
the population mean difference which comes from the
null hypothesis, and (2) the value for the pooled
The problem is that you cannot determine which of
these two values is responsible for a t statistic that
falls in the critical region.
In particular, you cannot be certain that rejecting the
null hypothesis is correct when you obtain an
extreme value for t.
Homogeneity of Variance Assumption
If the two sample variances appear to be
substantially different, you should use Hartley’s
F-max test to determine whether or not the
homogeneity assumption is satisfied.
 If homogeneity of variance is violated, Box
10.2 presents an alternative procedure for
computing the t statistic that does not involve
pooling the two sample variances.
Homogeneity of Variance Assumption
Hartley’s F-max test
Step 1: Compute Sample Variance for each separate sample

NOTE: If the F-max test rejects the
(s2 = ).

hypothesis of equal variances, or if
you suspect that the homogeneity of
variance assumption is not justified,
you should not compute an
independent-measures t-statistic
using pooled variance. In such
cases, use the alternative formula
for the t-statistic on the next slide.
Step 2: Compute F-Max =
Step 3: Find critical F-Max value
 k = number of separate samples (2)
 df = n – 1 for each sample (Hartley test assumes all
samples are the same size)
 Select an alpha (α) level for critical F-Max value - homogeneity tests use larger alpha levels at 0.05 or 0.01
Step 4: If computed F-Max < F-crit, then no evidence that the
homogeneity of variance assumption has been violated.
Homogeneity of Variance Assumption
Alternative formula for t-statistic
Step 1: Calculate the standard error using the two separate
sample variances as in equation 10.1
NOTE: The following alternative
formula for the t-statistic does not
pool sample variances and does
not require the homogeneity of
variance assumption.
Step 2: The value of degrees of freedom
for the t-statistic is adjusted using the
following equation:
Decimal values for df should be
rounded down to the next lower
integer. By lowering the df, you
push the boundaries of the critical
region farther out. This makes the
test more demanding.

similar documents