Hypothesis Test: Difference between revisions

From Rice Wiki
No edit summary
No edit summary
 
(7 intermediate revisions by the same user not shown)
Line 4: Line 4:
claim.
claim.


= Procedure =
This page will focus on ''usage'' of hypothesis tests ''in the context
of mean comparison''.


First, we need two things:
= Assumptions for HT =
 
We use the same assumptions used for a confidence interval for the
difference in means
 
# A random sample was taken
# Population 1 and 2 are independent
# Sample means are normally distributed
** Sample size is large enough (> 30)
** Population is normally distributed
 
= Procedure (Mean Comparison) =
 
== 1. Null and Alternative Hypothesis ==
 
To perform hypothesis test with mean comparison, we need two things:


* The '''null hypothesis <math>H_0</math>''' is the statement which we assume to be ''true''
* The '''null hypothesis <math>H_0</math>''' is the statement which we assume to be ''true''
* The '''alternative hypothesis <math>H_A</math>''' is the complement of the null hypothesis.
* The '''alternative hypothesis <math>H_A</math>''' is the complement of the null hypothesis.


Because sample mean has normal distribution, by RV linear combination,
Mean comparison work with the difference in means
the sampling distribution of <math>\bar{Y}_1 - \bar{Y}_2</math> is
 
<math>
\mu_1 - \mu_2
</math>
 
As such, there are three sets of hypotheses:
* <math>H_0: \mu_1 - \mu_2 = 0</math> vs <math>H_A: \mu_1 - \mu_2 \neq 0</math>
* <math>H_0: \mu_1 - \mu_2 \geq 0</math> vs <math>H_A: \mu_1 - \mu_2 < 0</math>
* <math>H_0: \mu_1 - \mu_2 \leq 0</math> vs <math>H_A: \mu_1 - \mu_2 > 0</math>
 
== 2. Test-Statistic ==
 
Next, we need to calculate a '''test-statistic <math>t_s</math>'''. This
measures how much our sample data differ from <math>H_0</math>. It
summarizes our data to one number to perform hypothesis test on.
 
For mean comparison, the hypothesized difference is 0 (i.e. the means
are the same). Therefore, the test-statistic is calculated as follows:
 
<math>
t_s = \frac{\bar{y_1} - \bar{y_2} - 0 }{ \sqrt{\frac{s_1^2}{n_1} + \frac{s_2^2}{n_2}}}
</math>
 
On the numerator, the difference mean is subtracted by 0 to since that is the
comparison point; all three sets of hypotheses in mean comparison
compares against 0.
 
On the denominator, the value is divided by the sample standard
deviation. This is a surprise tool that will help us later (bottom of
3.)
 
<math>t_s</math>, the more our data differs from <math>H_0</math>.
Notice that it increases with sample mean difference and decreases with
variance.
 
 
== 3. Find P-value ==
 
The '''p-value''' is the probability of observing our data or more
extreme if <math>H_0</math> is in fact true. To find this, we first need
to know the sampling distribution of our random variable.
 
=== Distribution ===
 
In the case of mean comparison, because sample mean has normal
distribution, by RV linear combination, the sampling distribution of
<math>\bar{Y}_1 - \bar{Y}_2</math> is


<math>
<math>
Line 18: Line 80:
</math>
</math>


We are not going to derive it, but the degree of freedom for this particular combination is
Both means follow the t-distribution, therefore the difference also
follows t-distribution.
 
We are not going to derive it, but the degree of freedom in this case is


<math>
<math>
v = \frac{ (\frac{s_1^2}{n_1} + \frac{s_2^2}{n_2} )^2 }
df = \upsilon = \frac{ (\frac{s_1^2}{n_1} + \frac{s_2^2}{n_2} )^2 }
{ \frac{(s_1^2 / n_1)^2}{ n_1 - 1} + \frac{(s_2^2 / n_2)^2}{ n_2 - 1} }
{ \frac{(s_1^2 / n_1)^2}{ n_1 - 1} + \frac{(s_2^2 / n_2)^2}{ n_2 - 1} }
</math>
</math>


'''Round down''' the value to use t-table.
where <math>\upsilon</math> is '''rounded down''' when using the t-table.
 
Remember how the test-statistic has the sample deviation on the
denominator? This is so that we can use the t-distribution to calculate
the probability! Now that we know the ''degrees of freedom'' and the
''test-statistic'' to compare against, we can calculate the p-value.
 
=== P-value ===
 
In the case of mean comparison, we have the following p-values:
 
* For <math>H_A: \mu_1 - \mu_2 \neq 0</math>, the p-value is <math>2P(t > |t_s|)</math>
** Two tails
* For <math>H_A: \mu_1 - \mu_2 > 0</math>, the p-value is <math>P(t > t_s)</math>
** Upper tail
* For <math>H_A: \mu_1 - \mu_2 < 0</math>, the p-value is <math>P(t < t_s)</math>
** Lower tail
 
The smaller the p-value, the less likely it is to observe our data or
more extreme if <math>H_0</math> is true, meaning that our data is
unlikely if our claim is true.
 
If there are no calculators, we can use a ''range'' of t-values that
contain the test-statistic. The result would be a lower and upper bound
p-value.
 
== 4. Conclusion ==
 
We decide a cutoff point for our p-values, typically at <math>\alpha =
0.1, 0.05, 0.01</math>, called the '''level of significance'''.
 
If <math>p < \alpha</math>, our data supports <math>H_A</math>,
therefore <math>H_0</math> is rejected. Otherwise, we failed to reject
<math>H_0</math>.
 


A CI that covers <math>0</math> implies that there is no significant
A CI that covers <math>0</math> implies that there is no significant
difference, as it is plausible for the population means to be equal.
difference, as it is plausible for the population means to be equal.
= Errors =
There are two types of possible errors for our conclusion.
* '''Type I error:''' Null is true but we rejected it
* '''Type II error:''' Null is false but we failed to reject it
Remeber the ''level of significance?'' Well it is actually
<math>
\alpha = P(Type I Error) = P(Reject H_0 | H_0 True)
</math>
So when setting the level of significance, we are actually controlling
type I error. The reason is that the distribution of <math>t_s</math> is
''non-specific:'' there are lots of possibilities that <math>H_0</math>
is false.
We also have
<math>
\beta = P(Type II Error) = P(FTR H_0 | H_0 False)
</math>
As <math>\alpha</math> increase, <math>\beta</math> decreases.
Therefore, we can choose what <math>\alpha</math> to use depending on
which error is probably worse. We ''cannot'' control <math>\beta</math>.
I don't really get why.
If we think Type I error is use, choose smaller <math>\alpha =
0.01</math>. Otherwise, use <math>\alpha = 0.10</math>.


[[Category:Sample Statistics]]
[[Category:Sample Statistics]]

Latest revision as of 01:55, 16 March 2024

Hypothesis test is a technique where sample data is used to determine if the confidence interval supports a particular claim. Hypothesis tests quantify how likely our data is given a particular claim.

This page will focus on usage of hypothesis tests in the context of mean comparison.

Assumptions for HT

We use the same assumptions used for a confidence interval for the difference in means

  1. A random sample was taken
  2. Population 1 and 2 are independent
  3. Sample means are normally distributed
    • Sample size is large enough (> 30)
    • Population is normally distributed

Procedure (Mean Comparison)

1. Null and Alternative Hypothesis

To perform hypothesis test with mean comparison, we need two things:

  • The null hypothesis is the statement which we assume to be true
  • The alternative hypothesis is the complement of the null hypothesis.

Mean comparison work with the difference in means

As such, there are three sets of hypotheses:

  • vs
  • vs
  • vs

2. Test-Statistic

Next, we need to calculate a test-statistic . This measures how much our sample data differ from . It summarizes our data to one number to perform hypothesis test on.

For mean comparison, the hypothesized difference is 0 (i.e. the means are the same). Therefore, the test-statistic is calculated as follows:

On the numerator, the difference mean is subtracted by 0 to since that is the comparison point; all three sets of hypotheses in mean comparison compares against 0.

On the denominator, the value is divided by the sample standard deviation. This is a surprise tool that will help us later (bottom of 3.)

, the more our data differs from . Notice that it increases with sample mean difference and decreases with variance.


3. Find P-value

The p-value is the probability of observing our data or more extreme if is in fact true. To find this, we first need to know the sampling distribution of our random variable.

Distribution

In the case of mean comparison, because sample mean has normal distribution, by RV linear combination, the sampling distribution of is

Both means follow the t-distribution, therefore the difference also follows t-distribution.

We are not going to derive it, but the degree of freedom in this case is

where is rounded down when using the t-table.

Remember how the test-statistic has the sample deviation on the denominator? This is so that we can use the t-distribution to calculate the probability! Now that we know the degrees of freedom and the test-statistic to compare against, we can calculate the p-value.

P-value

In the case of mean comparison, we have the following p-values:

  • For , the p-value is
    • Two tails
  • For , the p-value is
    • Upper tail
  • For , the p-value is
    • Lower tail

The smaller the p-value, the less likely it is to observe our data or more extreme if is true, meaning that our data is unlikely if our claim is true.

If there are no calculators, we can use a range of t-values that contain the test-statistic. The result would be a lower and upper bound p-value.

4. Conclusion

We decide a cutoff point for our p-values, typically at , called the level of significance.

If , our data supports , therefore is rejected. Otherwise, we failed to reject .


A CI that covers implies that there is no significant difference, as it is plausible for the population means to be equal.

Errors

There are two types of possible errors for our conclusion.

  • Type I error: Null is true but we rejected it
  • Type II error: Null is false but we failed to reject it

Remeber the level of significance? Well it is actually

So when setting the level of significance, we are actually controlling type I error. The reason is that the distribution of is non-specific: there are lots of possibilities that is false.

We also have

As increase, decreases. Therefore, we can choose what to use depending on which error is probably worse. We cannot control . I don't really get why.

If we think Type I error is use, choose smaller . Otherwise, use .