100% found this document useful (1 vote)
149 views

Hypothesis of Testing

1) A statistical hypothesis test involves stating a null hypothesis and an alternative hypothesis, formulating a test plan to analyze sample data, analyzing the sample, and interpreting the results to either reject or fail to reject the null hypothesis. 2) There are two types of errors in hypothesis testing - Type I errors occur when a true null hypothesis is rejected, while Type II errors occur when a false null hypothesis is not rejected. 3) Hypothesis test results are typically interpreted using either p-values, which measure the strength of evidence for the null hypothesis, or regions of acceptance/rejection defined by the significance level.

Uploaded by

faast_track
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
149 views

Hypothesis of Testing

1) A statistical hypothesis test involves stating a null hypothesis and an alternative hypothesis, formulating a test plan to analyze sample data, analyzing the sample, and interpreting the results to either reject or fail to reject the null hypothesis. 2) There are two types of errors in hypothesis testing - Type I errors occur when a true null hypothesis is rejected, while Type II errors occur when a false null hypothesis is not rejected. 3) Hypothesis test results are typically interpreted using either p-values, which measure the strength of evidence for the null hypothesis, or regions of acceptance/rejection defined by the significance level.

Uploaded by

faast_track
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 2

Statistics Tutorial: Hypothesis Tests

A statistical hypothesis is an assumption about a population parameter. This assumption may or


may not be true.
The best way to determine whether a statistical hypothesis is true would be to examine the entire
population. Since that is often impractical, researchers typically examine a random sample from
the population. If sample data are not consistent with the statistical hypothesis, the hypothesis is
rejected.
There are two types of statistical hypotheses.
 Null hypothesis. The null hypothesis, denoted by H0, is usually the hypothesis that sample
observations result purely from chance. 
 Alternative hypothesis. The alternative hypothesis, denoted by H 1 or Ha, is the hypothesis
that sample observations are influenced by some non-random cause.
For example, suppose we wanted to determine whether a coin was fair and balanced. A null
hypothesis might be that half the flips would result in Heads and half, in Tails. The alternative
hypothesis might be that the number of Heads and Tails would be very different. Symbolically,
these hypotheses would be expressed as
H0: P = 0.5 
Ha: P ≠ 0.5
Suppose we flipped the coin 50 times, resulting in 40 Heads and 10 Tails. Given this result, we
would be inclined to reject the null hypothesis. We would conclude, based on the evidence, that
the coin was probably not fair and balanced.
Can We Accept the Null Hypothesis?
Some researchers say that a hypothesis test can have one of two outcomes: you accept the null
hypothesis or you reject the null hypothesis. Many statisticians, however, take issue with the notion
of "accepting the null hypothesis." Instead, they say: you reject the null hypothesis or you fail to
reject the null hypothesis.
Why the distinction between "acceptance" and "failure to reject?" Acceptance implies that the null
hypothesis is true. Failure to reject implies that the data are not sufficiently persuasive for us to
prefer the alternative hypothesis over the null hypothesis.

Hypothesis Tests
Statisticians follow a formal process to determine whether to reject a null hypothesis, based on
sample data. This process, called hypothesis testing, consists of four steps.
 State the hypotheses. This involves stating the null and alternative hypotheses. The
hypotheses are stated in such a way that they are mutually exclusive. That is, if one is
true, the other must be false. 
 Formulate an analysis plan. The analysis plan describes how to use sample data to evaluate
the null hypothesis. The evaluation often focuses around a single test statistic. 
 Analyze sample data. Find the value of the test statistic (mean score, proportion, t-score,
z-score, etc.) described in the analysis plan. 
 Interpret results. Apply the decision rule described in the analysis plan. If the value of the
test statistic is unlikely, based on the null hypothesis, reject the null hypothesis. 
Decision Errors
Two types of errors can result from a hypothesis test.
 Type I error. A Type I error occurs when the researcher rejects a null hypothesis when it is
true. The probability of committing a Type I error is called the significance level. This
probability is also called alpha, and is often denoted by α.
 Type II error. A Type II error occurs when the researcher fails to reject a null hypothesis
that is false. The probability of committing a Type II error is called Beta, and is often
denoted by β. The probability of not committing a Type II error is called the Power of the
test.
Decision Rules
The analysis plan includes decision rules for rejecting the null hypothesis. In practice, statisticians
describe these decision rules in two ways - with reference to a P-value or with reference to a
region of acceptance.
 P-value. The strength of evidence in support of a null hypothesis is measured by the P-
value. Suppose the test statistic is equal to S. The P-value is the probability of observing a
test statistic as extreme as S, assuming the null hypotheis is true. If the P-value is less than
the significance level, we reject the null hypothesis.
 Region of acceptance. The region of acceptance is a range of values. If the test statistic
falls within the region of acceptance, the null hypothesis is not rejected. The region of
acceptance is defined so that the chance of making a Type I error is equal to the
significance level.
The set of values outside the region of acceptance is called the region of rejection. If the
test statistic falls within the region of rejection, the null hypothesis is rejected. In such
cases, we say that the hypothesis has been rejected at the α level of significance.
These approaches are equivalent. Some statistics texts use the P-value approach; others use the
region of acceptance approach. In subsequent lessons, this tutorial will present examples that
illustrate each approach.
One-Tailed and Two-Tailed Tests
A test of a statistical hypothesis, where the region of rejection is on only one side of the sampling
distribution, is called a one-tailed test. For example, suppose the null hypothesis states that the
mean is less than or equal to 10. The alternative hypothesis would be that the mean is greater than
10. The region of rejection would consist of a range of numbers located located on the right side of
sampling distribution; that is, a set of numbers greater than 10.
A test of a statistical hypothesis, where the region of rejection is on both sides of the sampling
distribution, is called a two-tailed test. For example, suppose the null hypothesis states that the
mean is equal to 10. The alternative hypothesis would be that the mean is less than 10 or greater
than 10. The region of rejection would consist of a range of numbers located located on both sides
of sampling distribution; that is, the region of rejection would consist partly of numbers that were
less than 10 and partly of numbers that were greater than 10.

You might also like