Professional Documents
Culture Documents
ANS:
statistical hypothesis:
It is a hypothesis that is testable on the basis of observed data modeled as the realised values
taken by a collection of random variables A set of data (or several sets of data, taken together)
are modelled as being realised values of a collection of random variables having a joint
probability distribution in some set of possible joint distributions. The hypothesis being tested is
exactly that set of possible probability distributions. A statistical hypothesis test is a method
of statistical inference. An alternative hypothesis is proposed for the probability distribution of
the data, either explicitly or only informally. The comparison of the two models is
deemed statistically significant if, according to a threshold probability -- the significance level --
the data is very unlikely to have occurred under the null hypothesis. A hypothesis test specifies
which outcomes of a study may lead to a rejection of the null hypothesis at a pre-specified level
of significance, while using a pre-chosen measure of deviation from that hypothesis (the test
statistic, or goodness-of-fit measure). The pre-chosen level of significance is the maximal
allowed "false positive rate". One wants to control the risk of incorrectly rejecting a true null
hypothesis.
Importance:
Statistics are helpful in analyzing most collections of data. This is equally true of hypothesis
testing which can justify conclusions even when no scientific theory exists. In the Lady tasting
tea example, it was "obvious" that no difference existed between (milk poured into tea) and (tea
poured into milk). The data contradicted the "obvious".
Real world applications of hypothesis testing include:
The null hypothesis can be thought of as the opposite of the "guess" the research made (in
this example the biologist thinks the plant height will be different for the fertilizers). So
the null would be that there will be no difference among the groups of plants. Specifically
in more statistical language the null for an ANOVA is that the means are the same. We
state the Null hypothesis as:
The reason we state the alternative hypothesis this way is that if the Null is rejected, there
are many possibilities.
If we look at what can happen in a hypothesis test, we can construct the following
contingency table:
In Reality
Decision \(H_0\) is TRUE \(H_0\) is FAL
Accept \(H_0\) OK Type II Error
β = probability
Reject \(H_0\) Type I Error OK
\(\alpha\) = probability of Type I Error
You should be familiar with type I and type II errors from your introductory course. It is
important to note that we want to set \(\alpha\) before the experiment (a-priori) because
the Type I error is the more ‘grevious’ error to make. The typical value of \(\alpha\) is
0.05, establishing a 95% confidence level. For this course we will assume \
(\alpha\) =0.05.
For categorical treatment level means, we use an F statistic, named after R.A. Fisher. We
will explore the mechanics of computing the F statistic beginning in Lesson 2.
The F value we get from the data is labeled \(F_{\text{calculated}}\).
As with all other test statistics, a threshold (critical) value of F is established.
This F value can be obtained from statistical tables and is referred to as \
(F_{\text{critical}}\) or \(F_\alpha\). As a reminder, this critical value is the minimum
value for the test statistic (in this case the F test) for us to be able to reject the null.