Type I Error (): This occurs when the null hypothesis () is actually true, but the statistical test provides enough evidence to reject it. It is often referred to as a false positive, similar to a medical test indicating a disease is present when the patient is actually healthy.
Type II Error (): This occurs when the null hypothesis () is false, but the test fails to provide sufficient evidence to reject it. This is known as a false negative, comparable to a medical test failing to detect a disease that a patient actually has.
Significance Level (): This is the pre-determined probability of committing a Type I error. By setting , a researcher accepts a risk of rejecting a true null hypothesis.
Power of a Test (): This represents the probability of correctly rejecting a false null hypothesis. A high power indicates a high likelihood of detecting an effect if one truly exists.
The Inverse Relationship: For a fixed sample size, decreasing the probability of a Type I error () inherently increases the probability of a Type II error (). This occurs because moving the decision threshold to be more conservative makes it harder to reject the null, even when it is false.
Probability Distributions: Errors are visualized as areas under probability density functions. is the area in the tail of the distribution, while is the area of the distribution that falls within the non-rejection region of .
Effect of Sample Size: Increasing the sample size () narrows the distributions (reduces standard error), which allows both and to decrease simultaneously. This is the primary method for improving the overall accuracy of a test.
Setting the Significance Level: Researchers must choose based on the consequences of a false positive. In high-stakes scenarios like criminal trials, is set very low (e.g., ) to avoid convicting the innocent.
Calculating Power: To find the probability of a Type II error, one must first define a specific alternative value for the parameter. is then calculated as .
Balancing Risks: A step-by-step approach involves: 1) Identifying the costs of each error type, 2) Selecting an appropriate , 3) Determining the required sample size to achieve a desired power ().
| Feature | Type I Error () | Type II Error () |
|---|---|---|
| Definition | Rejecting a true | Failing to reject a false |
| Nickname | False Positive | False Negative |
| Controlled By | Significance level () | Sample size and effect size |
| Consequence | Unnecessary action taken | Opportunity for action missed |
| Probability | Equal to | Calculated based on |
Contextual Identification: When asked to describe an error in context, always use the format: "The test concludes [Alternative Hypothesis] when in reality [Null Hypothesis] is true." This ensures you address both the decision and the reality.
The 'Accept' Trap: Never use the word "accept" for the null hypothesis. If you fail to reject , it does not mean is true; it simply means you lack evidence to prove it false. This is a common source of Type II error confusion.
Check the Relationship: If a question asks what happens to when is decreased, the answer is always that increases (assuming is constant).
Power and Beta: Remember that . If you are given the power of a test, you can immediately find the probability of a Type II error.
Misinterpreting : A common mistake is thinking is the probability that the null hypothesis is true. In reality, is a conditional probability: the probability of rejecting given that it is true.
Sum of Errors: Students often mistakenly believe that . This is incorrect because they are probabilities calculated under two different, mutually exclusive assumptions (one assuming is true, the other assuming is true).
Ignoring Effect Size: A Type II error is more likely when the true parameter is very close to the null value. The smaller the "effect size," the harder it is for a test to distinguish between the two distributions.