

Study with the several resources on Docsity
Earn points by helping other students or get them with a premium plan
Prepare for your exams
Study with the several resources on Docsity
Earn points to download
Earn points by helping other students or get them with a premium plan
Community
Ask the community for help and clear up your study doubts
Discover the best universities in your country according to Docsity users
Free resources
Download our free guides on studying techniques, anxiety management strategies, and thesis advice from Docsity tutors
Material Type: Exam; Class: Probability and Statistics; Subject: Economics; University: Vassar College; Term: Unknown 1989;
Typology: Exams
1 / 2
This page cannot be seen from the preview
Don't miss anything!
Statistical reasoning is a refinement of every day logical reasoning. Consider the statement, “If A , then B”. Lets suppose that this is a true statement. And let’s suppose that we wanted to investigate whether A was true. The knowledge that B is true does not provide us with evidence that A is true but the knowledge that B is false does provide us with evidence that A is false. If B is false we know that A is false but if B is true we don’t know anything about the truth or falsity of A.
Consider the following example. Suppose that the statement “all criminals have blonde hair” were true. We can rewrite this statement in the form “ If X is a criminal, then X has blonde hair” where X is a person. Clearly if X does not have blonde hair we know that X is not a criminal, but if X does have blonde hair we do not know whether or not X is a criminal. If we think of the statement “X is a criminal” as a hypothesis, then if X does not have blonde hair we know that the hypothesis is false and we say that we “reject the hypothesis”. If X has blonde hair however we don’t know that X is a criminal so we do not say that we “accept the hypothesis” that X is a criminal. We say instead that we “fail to reject the hypothesis” that X is a criminal. The same kind of reasoning applies to statistical hypothesis testing.
Example I. Suppose we want to test the hypothesis that a coin is a fair coin. (A fair coin is one for which the probability of getting a head is one-half.) In order to test this hypothesis we might choose to flip the coin 100 times. Our “if A, then B” statement becomes the following: If the coin is fair, then the number of heads which occur when we flip the coin 100 times is a normally distributed random variable with mean 50 and variance 25. Suppose that we flip the coin 100 times. (If the coin is fair this is equivalent to a random draw from a normal population with mean 50 and variance 25.) Suppose we get 70 heads. It would be extraordinarily unlikely that we would get this result if the population is normal with mean 50 and variance 25. So we reject the hypothesis that the coin is fair. If on the other hand we got 49 heads this is not an unusual outcome from a normal population with mean 50 and variance 25 we would fail to reject the hypothesis that the coin is fair.
Now you will notice a difference between this example and the former example. In the former example we assumed that we can say with certainty that “X does not have blonde hair.” In this example we can only say that it would be extraordinarily unlikely that we would get this result from a population normally distributed with mean 50 and variance 25.
Briefly, hypothesis testing consists of the following. A hypothesis is formulated. We call the hypothesis to be tested the “null” hypothesis. A experiment is run or some data is collected and a test statistic is calculated. Then the following question is asked.: Assuming that the hypothesis is true is the test statistic which we calculated unusual? If the test statistic is sufficiently unusual we conclude that the hypothesis is false and we “reject the hypothesis”. If the test statistic is not sufficiently unusual we “fail to reject the hypothesis.
Example II. Suppose that we have a population with unknown mean and standard deviation 10. Suppose that we want to test the hypothesis that the population mean is 100. We take a sample of size n = 25 and find a sample mean of 110. If the null hypothesis were true then we know
that and If then z = 5. This is a very unusual
result for a standard normal random variable. The probability of getting a result that differs from 0 by
this much is .000000574, or 1 chance in 1,742,169. So we have a choice. We can conclude that the
hypothesis that is incorrect or we can conclude that something with a probability of
.000000574 just occurred. Ordinarily we would choose the former explanation and reject the hypothesis that the population mean is 100. If, on the other hand, we got a sample mean of 98 we would conclude that this result is not very unusual from a population with mean 100, so we fail to reject the hypothesis.
Hypothesis testing can be carried out in distinct steps :
Step I. State the null hypothesis Step II. State the alternative hypothesis. Step III. Choose a test statistic Step IV. Calculate the distribution of the test statistic under the assumption that the null hypothesis is true. Step V Choose a significance level,. (Sometimes we skip this step and go straight to VI A.) Step VI (Do A or B)
A. Run the experiment and calculate the p-value. If the p-value is less than the significance level reject the null hypothesis. Otherwise fail to reject the null hypothesis. If we did not choose a significance level in step V we simply state the p-value.
B. Find a critical value of the test statistic. Run the experiment and calculate the value of the test statistic. Compare the calculated value of the test statistic with the critical value.
Type I and Type II error. There are two type of errors which you can make in a hypothesis test. A type I error is the error of rejecting the null hypothesis when it is true. A type II error is the error of failing to reject the null hypothesis when it is false. The significance level of a test is the conditional probability of rejecting the null hypothesis when the null hypothesis is true.
In general, if we increase the significance level we lower the probability of a type II error. When we choose a significance level for a test we have to consider the costs of making a type I or a type II error and the probability that the null hypothesis is true or false. Let C = the probable cost of an error. Then
where and are respectively the costs of a type I and a type II error.