Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

Hypothesis Testing - Exam Study Guide | ECON 209, Exams of Probability and Statistics

Material Type: Exam; Class: Probability and Statistics; Subject: Economics; University: Vassar College; Term: Unknown 1989;

Typology: Exams

Pre 2010

Uploaded on 08/16/2009

koofers-user-lor
koofers-user-lor 🇺🇸

10 documents

1 / 2

Toggle sidebar

This page cannot be seen from the preview

Don't miss anything!

bg1
Econ 209
Handout Hypothesis Testing
Statistical reasoning is a refinement of every day logical reasoning. Consider the sta tement, “ If
A , then B”. Lets suppose that this is a true statement. And let’s suppose that we wanted to investigate
whether A was true. The knowledge that B is true does not provide us with evidence that A is true but
the knowledge that B is false does provide us with evidence that A is false. If B is false we know that A
is false but if B is true we don’t know anything about the truth or falsity of A.
Consider the following example. Suppose that the statement “all criminals have blonde hair”
were true. We can rewrite this statement in the form “ If X is a criminal, then X has blonde hair” where
X is a person. Clearly if X does not have blonde hair we know that X is not a criminal, but if X does
have blonde hair we do not know whether or not X is a criminal. If we think of the statement “X is a
criminal” as a hypothesis, then if X does not have blonde hair we know that the hypothesis is false and
we say that we “reject the hypothesis”. If X has blonde hair however we don’t know that X is a criminal
so we do not say that we “accept the hypothesis” that X is a criminal. We say instead that we “fail to
reject the hypothesis” that X is a criminal. The same kind of reasoning applies to statistical hypothesis
testing.
Example I. Suppose we want to test the hypothesis that a coin is a fair coin. (A fair coin is one
for which the probability of getting a head is one-half.) In order to test this hypothesis we might choose
to flip the coin 100 times. Our “if A, then B” statement becomes the following: If the coin is fair, then
the number of heads which occur when we flip the coin 100 times is a normally distributed random
variable with mean 50 and variance 25. Suppose that we flip the coin 100 times. (If the coin is fair this
is equivalent to a random draw from a normal population with mean 50 and variance 25.) Suppose we
get 70 heads. It would be extraordinarily unlikely that we would get this result if the population is
normal with mean 50 and variance 25. So we reject the hypothesis that the coin is fair. If on the other
hand we got 49 heads this is not an unusual outcome from a normal population with mean 50 and
variance 25 we would fail to reject the hypothesis that the coin is fair.
Now you will notice a difference between this example and the former example. In the former
example we assumed that we can say with certainty that “X does not have blonde hair.” In this example
we can only say that it would be extraordinarily unlikely that we would get this result from a population
normally distributed with mean 50 and variance 25.
Briefly, hypothesis testing consists of the following. A hypothesis is formulated. We call the
hypothesis to be tested the “null” hypothesis. A experiment is run or some data is collected and a test
statistic is calculated. Then the following question is asked.: Assuming that the hypothesis is true is the
test statistic which we calculated unusual? If the test statistic is sufficiently unusual we conclude that the
hypothesis is false and we “reject the hypothesis”. If the test statistic is not sufficiently unusual we “fail
to reject the hypothesis.
Example II. Suppose that we have a population with unknown mean and standard deviation 10.
Suppose that we want to test the hypothesis that the population mean is 100. We take a sample of size n
= 25 and find a sample mean of 110. If the null hypothesis were true then we know
that and If then z = 5. This is a very unusual
result for a standard normal random variable. The probability of getting a result that differs from 0 by
pf2

Partial preview of the text

Download Hypothesis Testing - Exam Study Guide | ECON 209 and more Exams Probability and Statistics in PDF only on Docsity!

Econ 209

Handout Hypothesis Testing

Statistical reasoning is a refinement of every day logical reasoning. Consider the statement, “If A , then B”. Lets suppose that this is a true statement. And let’s suppose that we wanted to investigate whether A was true. The knowledge that B is true does not provide us with evidence that A is true but the knowledge that B is false does provide us with evidence that A is false. If B is false we know that A is false but if B is true we don’t know anything about the truth or falsity of A.

Consider the following example. Suppose that the statement “all criminals have blonde hair” were true. We can rewrite this statement in the form “ If X is a criminal, then X has blonde hair” where X is a person. Clearly if X does not have blonde hair we know that X is not a criminal, but if X does have blonde hair we do not know whether or not X is a criminal. If we think of the statement “X is a criminal” as a hypothesis, then if X does not have blonde hair we know that the hypothesis is false and we say that we “reject the hypothesis”. If X has blonde hair however we don’t know that X is a criminal so we do not say that we “accept the hypothesis” that X is a criminal. We say instead that we “fail to reject the hypothesis” that X is a criminal. The same kind of reasoning applies to statistical hypothesis testing.

Example I. Suppose we want to test the hypothesis that a coin is a fair coin. (A fair coin is one for which the probability of getting a head is one-half.) In order to test this hypothesis we might choose to flip the coin 100 times. Our “if A, then B” statement becomes the following: If the coin is fair, then the number of heads which occur when we flip the coin 100 times is a normally distributed random variable with mean 50 and variance 25. Suppose that we flip the coin 100 times. (If the coin is fair this is equivalent to a random draw from a normal population with mean 50 and variance 25.) Suppose we get 70 heads. It would be extraordinarily unlikely that we would get this result if the population is normal with mean 50 and variance 25. So we reject the hypothesis that the coin is fair. If on the other hand we got 49 heads this is not an unusual outcome from a normal population with mean 50 and variance 25 we would fail to reject the hypothesis that the coin is fair.

Now you will notice a difference between this example and the former example. In the former example we assumed that we can say with certainty that “X does not have blonde hair.” In this example we can only say that it would be extraordinarily unlikely that we would get this result from a population normally distributed with mean 50 and variance 25.

Briefly, hypothesis testing consists of the following. A hypothesis is formulated. We call the hypothesis to be tested the “null” hypothesis. A experiment is run or some data is collected and a test statistic is calculated. Then the following question is asked.: Assuming that the hypothesis is true is the test statistic which we calculated unusual? If the test statistic is sufficiently unusual we conclude that the hypothesis is false and we “reject the hypothesis”. If the test statistic is not sufficiently unusual we “fail to reject the hypothesis.

Example II. Suppose that we have a population with unknown mean and standard deviation 10. Suppose that we want to test the hypothesis that the population mean is 100. We take a sample of size n = 25 and find a sample mean of 110. If the null hypothesis were true then we know

that and If then z = 5. This is a very unusual

result for a standard normal random variable. The probability of getting a result that differs from 0 by

this much is .000000574, or 1 chance in 1,742,169. So we have a choice. We can conclude that the

hypothesis that is incorrect or we can conclude that something with a probability of

.000000574 just occurred. Ordinarily we would choose the former explanation and reject the hypothesis that the population mean is 100. If, on the other hand, we got a sample mean of 98 we would conclude that this result is not very unusual from a population with mean 100, so we fail to reject the hypothesis.

Hypothesis testing can be carried out in distinct steps :

Step I. State the null hypothesis Step II. State the alternative hypothesis. Step III. Choose a test statistic Step IV. Calculate the distribution of the test statistic under the assumption that the null hypothesis is true. Step V Choose a significance level,. (Sometimes we skip this step and go straight to VI A.) Step VI (Do A or B)

A. Run the experiment and calculate the p-value. If the p-value is less than the significance level reject the null hypothesis. Otherwise fail to reject the null hypothesis. If we did not choose a significance level in step V we simply state the p-value.

B. Find a critical value of the test statistic. Run the experiment and calculate the value of the test statistic. Compare the calculated value of the test statistic with the critical value.

Type I and Type II error. There are two type of errors which you can make in a hypothesis test. A type I error is the error of rejecting the null hypothesis when it is true. A type II error is the error of failing to reject the null hypothesis when it is false. The significance level of a test is the conditional probability of rejecting the null hypothesis when the null hypothesis is true.

In general, if we increase the significance level we lower the probability of a type II error. When we choose a significance level for a test we have to consider the costs of making a type I or a type II error and the probability that the null hypothesis is true or false. Let C = the probable cost of an error. Then

where and are respectively the costs of a type I and a type II error.