Sampling Distribution: Difference between revisions
Line 57: | Line 57: | ||
= Binomial Normal Approximation = | = Binomial Normal Approximation = | ||
The sampling distribution of the | The sampling distribution of the proportion of success of ''n'' bernoulli random variables (<math>\hat{p}</math>) can also be approximated to a normal distribution under the CLT. | ||
Consider [[Discrete Random Variable#Bernoulli|bernoulli]] random variables | Consider [[Discrete Random Variable#Bernoulli|bernoulli]] random variables | ||
Line 63: | Line 63: | ||
<math>Y_1 \ldots Y_n \sim E(Y_i) = p, Var(Y_i) = p(1 - p)</math> | <math>Y_1 \ldots Y_n \sim E(Y_i) = p, Var(Y_i) = p(1 - p)</math> | ||
The | The proportion of success <math>\hat{p}</math> is the sum over the count, so the expected probability is | ||
<math>E(\hat{p}) = E \left(\frac{1}{n} \sum Y_i \right) = p</math> | <math>E(\hat{p}) = E \left(\frac{1}{n} \sum Y_i \right) = p</math> | ||
Line 72: | Line 72: | ||
With a large sample size ''n'', we can appoximate this to a normal distribution. Notably, the criteria for a large ''n'' is different from that of the continuous random variable. | With a large sample size ''n'', we can appoximate this to a normal distribution. Notably, the criteria for a large ''n'' is different from that of the continuous random variable. | ||
<math>np > 5, n(1 - p) > 5</math> | |||
then we have | |||
<math>\hat{p} \sim N \left( p, \frac{p(1-p)}{n} \right)</math> | |||
The reasoning behind the weird criteria relates to the [[Discrete Random Variable|binomial distribution]]. It's not very elaborated on in the lecture, but <math>np</math> is the mean of the binomial (i.e. the expected number of successes). The criteria essentially makes sure that no negative values are plausible in the approximation; with a small mean and a large variance, the left side of a normal approximation goes into the negative, but bernoulli/binomial should always be positive by common sense. | |||
= T-Distribution = | = T-Distribution = |
Revision as of 16:41, 19 March 2024
Let there be , where each is a randomv variable from the population.
Every Y have the same mean and distribution that we don't know.
We then have the sample mean
The sample mean is expected to be through a pretty easy direct proof
The variance of the sample mean is , also through a pretty easy direct proof.
Central limit theorem
The central limit theorem states that the distribution of the sample mean follows normal distribution.
As long as the following two conditions are satisfied, CLT applies, regardless of the population's distribution.
- The population distribution of is normal, or
- The sample size for each is large
By extension, we also have the distribution of the sum.
where
Confidence Interval
Estimation is the guess for the unknown parameter. A point estimate is a "best guess" of the population parameter, where as the confidence interval is the range of reasonable values that are intended to contain the parameter of interest with a certain degree of confidence, calculated with
(point estimate - margin of error, point estimate + margin of error)
Constructing CIs
By CLT, . The confidence interval is the range of plausible .
If we define the middle 90% to be plausible, to find the confidence interval, simply find the 5th and 95th percentile.
Generalized, if we want a confidence interval of the middle , have a confidence interval of
where is the sample mean and is the z score of the x-th percentile.
Binomial Normal Approximation
The sampling distribution of the proportion of success of n bernoulli random variables () can also be approximated to a normal distribution under the CLT.
Consider bernoulli random variables
The proportion of success is the sum over the count, so the expected probability is
and the variance of is
With a large sample size n, we can appoximate this to a normal distribution. Notably, the criteria for a large n is different from that of the continuous random variable.
then we have
The reasoning behind the weird criteria relates to the binomial distribution. It's not very elaborated on in the lecture, but is the mean of the binomial (i.e. the expected number of successes). The criteria essentially makes sure that no negative values are plausible in the approximation; with a small mean and a large variance, the left side of a normal approximation goes into the negative, but bernoulli/binomial should always be positive by common sense.
T-Distribution
CLT has several restrictions, the biggest one being a large sample size. T-
Since we don't know the population variance , we have to use the sample variance to estimate it. This introduces more uncertainty, accounted for by the t-distribution.
T-distribution is the distribution of sample mean based on population mean, sample variance and degrees of freedom (covered later). It looks very similar to normal distribution.
When the sample size is small, there is greater uncertainty in the estimates. T-di
The spread of t-distribution depends on the degrees of freedom, which is based on sample size. When looking up the table, round down df.
As the sample size increases, degrees of freedom increase, the spread of t-distribution decreases, and t-distribution approaches normal distribution.
Based on CLT and normal distribution, we had the confidence interval
Now, based on T-distribution, we have the CI
Find Sample Size
To calculate sample size needed depending on desired error margin and sample variance by assuming that
We want to always round up to stay within the error margin.
I don't really know why.
Sampling Distribution of Difference
By linear combination of RVs, sampling distribution of is
However, we do not know the population variance . If the CLT assumptions hold, then we have
Trust me bro. Remember to round down to use t-table. With this degree of freedom, we can use sample variance to estimate the distribution.