









Estude fácil! Tem muito documento disponível na Docsity
Ganhe pontos ajudando outros esrudantes ou compre um plano Premium
Prepare-se para as provas
Estude fácil! Tem muito documento disponível na Docsity
Prepare-se para as provas com trabalhos de outros alunos como você, aqui na Docsity
Os melhores documentos à venda: Trabalhos de alunos formados
Prepare-se com as videoaulas e exercícios resolvidos criados a partir da grade da sua Universidade
Responda perguntas de provas passadas e avalie sua preparação.
Ganhe pontos para baixar
Ganhe pontos ajudando outros esrudantes ou compre um plano Premium
Comunidade
Peça ajuda à comunidade e tire suas dúvidas relacionadas ao estudo
Descubra as melhores universidades em seu país de acordo com os usuários da Docsity
Guias grátis
Baixe gratuitamente nossos guias de estudo, métodos para diminuir a ansiedade, dicas de TCC preparadas pelos professores da Docsity
This chapter completes the preliminaries for studying stochastic processes. It discusses bernoulli processes, which are sums of independent and identically distributed random variables, and their properties. The chapter also covers the distribution of numbers of successes in bernoulli processes and the relationship between the times of successes and numbers of successes.
Tipologia: Notas de estudo
1 / 15
Esta página não é visível na pré-visualização
Não perca as partes importantes!
42 EXPECTATIONS AND INDEPENDENCE^ Chap. 2
Let N be a non-negative integer-valued random variable which is independent of {Xt,X 2 ,. ..}, andlet
E[N] = c, Var(AO = d^2.
Let S 0 = O, S^1! = Jfj, 5 2 = Jf, + JSf 2 ,.. ., and let Y = SN. (a) ComputeE[F|Aa E[YZ\N]. (b) Compute E[ Y], E[ Y^2 ], Var( Y). (c) Show that for any a e K+,
where
The chaptcr just finished completes our account of the preliminaries necessary for stuâying stochastic processes. Especially in vietv ofthe monstrous looks ofthe last section, it seerns ali too advisable to inquire how the reader's patience is holâing out and to assure him that he ivill in time come to appreciate the true frienáliness ofthese concepts. For a âeeper treatment andfor the proofs which wehave omitteâ, we refer the reader to CHUNG [2].
Consider an experiment consisting of an infinite sequence of trials. Sup- pose that the trials are independent of each other and that each one hás only two possible outcomes: "success" and "failure." A possible outcome of such an experiment is (S, F, F, S, F, S,.. .), which stands for the outcome where the first trial resulted in success, the second and third in failure, fourth in success, fifth in failure, sixth in success, etc. Thus, the sample space Q of the experiment consists of ali sequences of two letters 5" and F, that is,
Q = [a>: co = (coi, co 2 ,.. .), ca, is either S or F}.
We next describe a probability measure P on ali subsets of íl. Let O < p < l, and define q = l — p. We think of p and q as the probabilities of success and failure át any one trial. The probability of the event {ca: ol — S, a> 2 = S, o> 3 = F] should then be p-p-q — p^2 q, and the probability ofthe event {co: o>i = S, o) 2 = F, co 3 = F,a> 4 = F, co 5 = S} should be p-q-q-q-p = p^2 q^3. For each n, we consider ali events which are specified by the first n trials and define their probabilities in this manner. This, in addition to the conditions in Defini- tion (1.1.6), completely specifies the probability P. For each co e Í2 and n e {1,2,...}, define Xn(
•l!
BERNOULLI PROCESSES Chap. 3^ Sec.^2 NUMBERS OF SUCCESSES^47
\ N,,; n e N} whose time parameter is discrete N. = (O, l ,.. .} and whose state space is discrete and again is [O, l ,.. .}. We start with simple descriptiy_e quantities. Using the definition of Nn, Corollary (2.1.25), and the result (1.5) above, we obtain
E[Xn] = np
for ali n > l, and clearly, E[N 0 ] = O, só E[Na] = np for ali n > 0. Similarly, using the result in Exercise (2.3.9) and the independence of X^..., Xn,
by (1.6). From (2.2) and (2.3) we can get, by using the formula (2.1.28) for the variance,
(2.4) E[Nl\ = Var(AO + (E[Nn])^2 = npq + n
We could have computed E[N%] directly by noting that
só that
E[N
] = i; E[xn + Ê s£[M] ;=i /=i w = np + n(n — T)p^2 = n^2 p^2 + npq.
This last computation indicates how to go about obtaining the higher mo- ments of N,. Note that we have computed the expected value and variance without knowing the distribution of Nn. However, if one needs to compute E[g(Nn)] for arbitrary g, then we need the distribution of Nn. This we do next. The use of conditional probabilities in the proof below illustrates one of the main techniques found useful in studying stochastic processes. ^ ^ £*-
(2.5) LEMMA. For any w, A: e N *
P{Nn+1 = k} = pP{Nn = k-l} + qP{Nn = k}.
Proof. Fix n and k. Using Theorem (1.3.3) with A = {Na+l = k] and
/ ' l ^ (^) B , i = * } =A-} :^ - Ç^ P{Nn,, =k\Nn= j ] P{Nn^ = j }.
Since [Xlt... , Xn] is independent of Xn+1, by Definition (2.2.24), Nn =
P{Nn+1 = k\Nn=j} = P{Xn+í =
O otherwise.
Putting this in (2.6), we obtain the desired result. (^) D
For n = O, N 0 = O and therefore P{N 0 = 0} = l, and P{N 0 = k] = O for k 3=. 0. These, along with Lemma (2.5), specify the values of P{Nt = k}. The values of P{A^i = k} found, along with Lemma (2.5), give the values of P[N 2 = k], etc. Doing this a few steps, we obtain the table shown in Figure 3.2.1. From the table we recognize the nth row as the terms in the binomial expansion of (q + p)". Then we prove this conjecture.
n
. n n n n
= 0:
= 2: = 3: = 4:
1 q l^2 l^3 q*
k = l
0 P 2qp 3q^2 p W P
k = 2
Q 0 P^2 3qp^2 6qlp>
k = 3
0 0 0 P^3 4qp^3
k =
0 0 0 0 P
Figure 3.2.1 Using the probabilities P{Nn = k] in one row, one obtains the values P[Nn+ 1 = k] in the next row.
(2.7) THEOREM. For any n e N
P{Nn = (^) *, k = O,... , «.
Proof. For n = O, the claim is true, since NQ = 0. We now make the induction hypothesis that the formula is true for n = m and ali k. To complete the proof we need to show that, then, the claim is also true for n = m + l and ali k. For n = m + l, k = O, Lemma (2.5) and the induction hypothesis give
P{Nm+l =0) =p.Q + q.qn> = q™+i
as claimed. For n = m + l and O < k < m + l, again by Lemma (2.5) and
the induction hypthesis, we have
mi k(m- k)' ml (k- l)(m -k) (m+1) kl(m+ l - k)\l
só the claim is again true.
nkgm-k+l^ *^ i^ J_
p q \m - k + l + k.
D
We have now specified the distribution of the number of successes in the first n trials. Noting that Nm+n — Nm = Xm+l + • • • + Xm+n ajsa is the sum of n independent and identically distributed Bernoulli variables, we deduce that the distribution of Nm+n — Nm is the same as that of Nn. This we state as ~
(2.8) COROLLARY. For any m, n e
=
independent of m.fr
£. The distributfon figuring above is called the binomial distribution. As we ^ had noted before, P{Nn = k] is the fcth term in the binomial expansion of (q + p)". The coefficient of the &th term is usually denoted by ( j? j (read n above k), that is,
(29) ,k) k(n-k)\
(2.10) EXAMPLE. Using Corollary (2.8), we obtain
= 4} =
S<> lhe corollary above enables us to compute the distribution of N,,H.a — Nni l « ' r a l i n, in. I5ul we still are not able to compute compound probabilities
Sec. (^2) NUMBERS OF SUCCESSES (^) 49
such as P{N 5 = 4, N-, = 5, N 13 = 8). The means for doing that is furnished by the following
(2.11) THEOREM. For any m, n e N
P{Nm+n ~Nm = k\N 0 ,...,Nm} = P{N,n+n ~Nm = k} = (nk )^ <T *
for k = O,... , n.
(2.12) EXAMPLE. To illustrate the use of this theorem, suppose we are interested in evaluating the joint probability
The event whose probability is desired is equal to the event {N 5 = 4, N 7 — N 5 = l, N (^) Í 3 — N 7 = 3}. By Theorem (2.11), the random variable N 13 — N^ is independent of N 0 ,... , N 7. Thus,
P{7Y 5 =4,N 7 -N 5 = l, N 13 - N 7 = 3} = P{N 13 - N 7 = 3}P{N 5 =4,N 7 ~ N 5 = 1}.
Again by Theorem (2.11), N 7 — N 5 is independent of N „,..., N 5 and there- , • fore of N, in particular. Hence, ,. 'fi' ~~~~ " '- ..... -• -, ;.. • ;.,-.. '•! '
Putting ali these together,
= 8} = P[N 5 = 4}P{N 7 = 1}P{N 13 ~ (^393) = 200p»q',
where the last line followcd from Corollary (2.8), which was also put in Theorem (2. 11). Finally, suppose we want to compute E[NSN^]. Writing Ns = N 5 + (Ns — N 5 ), we have
E[N 5 NS] = E[Ns(Nf + (Ns - Ns))] = E[NI + Ns(Nt -N,)] = E[N^25 ] + E[N 5 ]E[NS - N 5 ],
where to get the last line we first used Corollary (2.1.25), and then used the independence of yV 5 and NS—NS (which is implied by Theorem (2. 1 1) above)
BERNOULLI PROCESSES Chap. 3
by Proposition (2.2.15). Putting the last two results in (2.15), we obtain
E[Y\ ATO,... , 7VJ = £ h(Nn, i,, ..., 0*0-,) • • • jr(O = f(Nm) i
independent of A^,... , Nm_í. Finally, by Corollary (2.2.22), this implies that
E[Y\N 0 , ...,Nm} = f(Nm) = E[Y\Nm],
as was to be shown. G
Theorem (2.14) above is satisfied by many processes of much less specific structure. Such processes are called Markov chains, and we will later study them at some length. The particular Markov chain we have, namely [Na; n e N], hás the further property mentioned in Corollary (2.13), and this makes our present job easier. The following is to illustrate the workings of Theorem (2.14).
(2.16) EXAMPLE. We want to compute the conditional expectation of Ni: given Ns. We have
The second equality made use of Corollary (2.2.12); to pass to the third equality we used the fact that E[NS \N 5 ] = N 5 by Proposition (2.2.15), and the fact that N, (^) t — N 5 is independent of N 5 by Corollary (2.13). G
(2.17) EXAMPLE. Let us compute E[N 5 Ng] again (cf. Example(2.12)forthe same) to show that conditional expectations, used properly, can reduce the work involved. By Proposition (2.2.16), E[N 5 Ns] = E[E[N,Ng\N 5 ]]. On the otherhand,
E[N 5 NS \ Ns] = N 5 (N 5 + 3p).
Hence,
3pN 5 ] = E[Nl] + 3pE[N 5 ] = 25p^2 + 5pq + 3p-5p = 40p^2 + 5pq.
(2.18) EXAMPLE. We want to compute
z E\N 5 Nn\N 2 ,N 3 ].
G
Sec. 3 TIMES OF SUCCESSES
By Theorems (2.2.19) and (2.14),
(2.19) Z = E[E[Nstf,, | tf 0 ,... , N 5 ] | tf 2 , N 3 ]
Using the result of Example (2.16),
E[N 5 Nl 11 tf 5 ] = N 5 £[tf, j | tfs] = N 5 (N 5 + só that
Again applying Theorems (2.2.19) and (2.14) and Proposition (2.2.15) in that order, Z = E[E[N^25 1 N 0 , N,, N 2 , N 3 ] \ N 2 , N 3 ]
(that is, given N 2 and N 3 , we need only N 3 as far as predicting the future is concerned). Writing N 5 = N 3 + (tf 3 — tf 3 ), thenusing Corollary (2.13) to deduce the independence of tf, — tf 3 and tf 3 , we get
Z = E[Nl + 2N,(tf 5 - tf,) + (tfs - tf 3 )^2 + 6pN 5 [ tf 3 ] = tf l + 2N 3 E[N 5 - N 3 1 tf 3 ] + E[(NS - tf 3 )^21 tf 3 ] + 6pE[N 5 \ N 3 ] = tf^2 -f 2N 3 E[N 5 - tf 3 ] + E[(NS - tf 3 )^2 ] + 6p(N 3 + 2p) == tf^2 + 2N 3 -2p + 4p^2 + 2^ + 6/?tf 3 + Up^2
Wc now have a number of theorems at our disposal to compute the prob- abililics or expectations rclatcd to the sequence {N,,}. In the next section we will obtain results o f a similar nalurc about the sequence of times (random) M! which the successes occui'.
'. '1'inies of Successes
l cl | A ' , , ; / / ' l \ l H" a Hcrninilli procirss w i t h probability/; of success. For l i x c i l <o • í i, considcr l h o i v a l í / ; i l i < m A' (^) t(m). \ ..(<»), A ',(<•;>),... ofthis process. I l u. i', a se<|iicnu' nl' oiu-s a i u l /rio:,. |)ciiolc hy '/',(o>), T 2 (co),... the n n l u i ". corresponding l « > l h e :,IHIT:,NÍVI' DiK's. l''or example, if X^at) = O, \ .(.->) l , A , ( o > ) d, A ',((„) I , A '. , ( (^) ( , < ) l Ihcn r,(co) = 2, T 2 (co)
j = k \ J /
V v^ * /
j k
58 BERNOULLI PROCESSES Cfldp. 3
by Corollary (3.10), 7, T 5 — TI, T 7 — T 5 are independent, we can write
P{T, = 3, r, - r, = 6, r 7 - r 5 = 8} = P{Tl = 3}P[T 5 -Tl= 6}P{T 7 - T 5 = 8} = P{T 1 = 3}P{T, = 6}P{T 2 = 8],
where the last equality is again obtained from (3.10). Now using Theorem (3.3) to evaluate these, we obtain
P{Tl = 3, T 5 = 9, T 7 = 17} = P{r, = 3}P{T 4 = 6}P{T 2 = 8}
= G»?^2 )(iQpVX7/»V) =^ D
(3.17) EXAMPLE. A certain component in a large system hás a lifetime whose distribution can be approximated by n(ni) = pqm~l, m > 1. When the component fails, it is replaced by an identical one. Let Tí, T 2 , • • • denote the times of failure; then Uk = Tk — TV ^ is the lifetime of the A:th item replaced. Since the components are identical,
and the items have independent lifetimes. Thus [Tk] can be looked upon as the times of "successes" in a Bernoulli process. Past records indicate that the first three failures occurred at times 3, 12, and 14. We would like to estimate the time of the fifth failure. By Theorem (3.6),
Writing T 5 = T 3 + (T, — T 3 ), using Corollary (2.2.12) to write the condi- tional expectation as the sum of the conditional expectations, noting that E[T 3 \ T 3 ] = T 3 by Proposition (2.2.15) and that E[T 5 - T 3 \ T 3 ] = E[T 5 — J 3 ] by the independence of T 5 — T 3 from T 3 (this is by Corollary (3.10)), we have
Hence, the quantity of interest is
E[T 5 \T 1 =3,T 2 = 12, T 3 = 14] = 14 + J-.
If/> = 0.60, this is 17.33. D
(3.18) EXAMPLE. In the preceding example suppose that each item costs c dollars when new. Assuming that a dollars today is worth l dollar during the next period, we now compute the expected total discounted value of ali future expenditures on this component. It is assumed that the discount rate remains the same throughout. There- fore, c dollars spent at time n is worth ca" dollars at present. Note that a is less than l ; if the "inflation" rate is 0.06, for example, then a = 1/(1 + 0.06). Fixco e Cl; the Mi replacement occurs at time Tk(ca). That kth item costs c dollars, and therefore its discounted value is
Thus, the total discounted cost is
co e £2. k=\
The expected value of a sum being the sum of the expectations (cf. Theo- rem (2.1.34)), we have
E[C\ = f] cE[uT*]. k=l
Writing Tk = T, + (T 2 - T,) + ---- h (Tk - TV,), noting that TlfT 2 - TI,... ,T.k — Tk^i are independent by Proposition (3.9), and applying Pro- position (2. l .26) on the expected value of such products, we obtain
E[y, -]
Since T,,T 2 — TI,... have the same geometric distribution,
Ik-ncc,
c (^) \ «e.ou/ , a/; c n
We close this section with a gcneralization of the computations appearing in l'xaniplc (3.18). Suppose that the value of the amount spent at time n is «•t|ii;il Io /'(") in tcrms of present worth. In the preceding example we had / ( / ; ) ca". Now we only assume that f (ri) > O for ali n. In fact, the result ( l l')) helow hokls for any / for which J] /(«) K^ absolutely convergent.
60 BERNOULLI^ PROCESSES^ Chap. 3
We would like to compute
Without further information onf, it looks as if we have to use Theorem (3.3). But we can do better if we examine the sum S = 2*=i f(Tk). If, for co e Q, r, (o) = 2, r 2 (eo) = 5, r 3 (<») = 10, r 4 (co) = 11, etc., then S(<a) = /(2) + /(5) + /(IO) + /(!!)+ • • •. Note that 2, 5, 10, 11,... are precisely those índices n for which Zn(co) = l, and that 5(to) = /(l) Z,(o) + /(2) Z 2 (co) + /(3) X 3 (co) + • • •. This is true in general, and we can write
The rest is easy :
= S f(n~)E[Xn] = p 2 /(«). n = l n = l
4. Sums of Independent Random Variables
In comparing the stochastic structures of the process [Nn; n e N] of Sec- tion 2 and the process {T,,; n e N] of Section 3, as revealed for example by Corollaries (2.13) and (3.10) respectively, we see that they are very much alike. As we mentioned before, this is due to the facts that N! — N 0 , N 2 — N I ,... are independent and identically distributed and that TI — T 0 ,T 2 — TI,... are also independent and identically distributed. It is this property that we now will put in a general setting. Let Ylt Y 2 ,. • • be a sequence of independent and identically distributed random variables, and define for each n e N
Yl
if n = O, i f « > 1.
We are interested in the stochastic process [Zn; n e N] and, especially, its limiting behavior as n approaches infinity. These limit theorems have applica- tions in data collection and statistical estimation, and we shall point out the way they are used. The following property is immediate from the definition of the Zn and Ddiiiilion (2.2.24) of indcpcndence.
SUMS OF INDEPENDENT RANDOM VARIABLES 61
(4.2) PROPOSITION. The stochastic process {Zn; H e N) hás stationary and independent increments; that is, (a) ZBI — ZBO, ZB2 — Zn i ,... , Znt — Znt^ are independent for any in- tegers O < HO < nl < • • • < nk\ and (b) the distribution of ZB+m — Zn is independent of n.
Conversely, if a process {Zn} hás stationary independent increments, then taking H 0 = O, H, = l,.. ., nk = k shows that Zt — Z 0 , Z 2 — Zj,... are independent and identically distributed. Hence, ali discrete parameter pro- cesses with stationary independent increments are obtained in this manner. Proposition (4.2) can be used to prove, just as in Theorem (2.14), the fol- lowing result: in words, as far as predicting the value of a random variable W depending on the future Zn, Zn + 1 ,... of the process {ZB} is concerned, allpast information concerning Z 0 ,... , ZB_t becomes worthless once the present value Za is given.
(4.3) PROPOSITION. Suppose W = g(Zn, Zn+1,.. .) for some function g > O and some integer n e N. Then
_E\Yr_^ CTM/I 7 j ^Q, • • • ,^ 7 1 _-in_^ — I-j\jV P\W\ 7 j ^rt^ l j'
The following computations are easily made by using the definition of Zn given by (4.1). Suppose
(4.4) E[Yn] = a, Var(Fn) = b\
Then (4.1) and Corollary (2.1.25) give
(4.5) E[Zn] = na,
and (4.1), the independence of Y1} Y 2 ,.. ., and (2.3.9) imply
(4.6) Var(Z,,) = nb^2.
(4.7) EXAMPLC. Successivc customcrs arriving at a gás station for fuel spcnd independent amounts. Lct Y{, K 2 ,... be the amounts (measured in ilollars) demanded by the lirst, second,... customers. Then Z,, is theamount ilcmandcd by the first n customers. If an avcrage demand is 4.87 dollars, Ilicn the amount demanded by the first 9 customers hás the expected value ') 4.87 = 43.83 dollars. G
In lhe preceding example it is easy to convince ourselves (by considering 11 u- iiulepcndcncc of the actions of different customers) that YÍ,Y 2 ,... are indi-pcndent and identically distributed. But how do we find out the expected
64 BERNOULLI PROCESSES Chap. 3
probability of success is
p = P{Xk = 1} = P{Yk <t} = (p(t).
Now, note that Fn(t, co) = Nn(co)/n is the average number of successes in the first n trials. Applying Theorem (4.10) to the process [Na], we get (as in Exam-
lim Fn(t, 03) = Hm :^ == p =
For fixed n and w, t — > Fn(í, co) is a step function increasing at each num- ber Yk(aj) by a jump of magnitude l /n. It is called the empirical distribution based on n observations. Proposition (4.13) shows that for large n it approxi- mates the distribution function ç?. We have seen in Proposition (4.9) that for large n the average ZJn is close to the expected value a of the increments Yk. Somewhat better information is contained in the following famous result. It is called the central limit theorem.
(4.14) THEOREM. For any t e K,
lim
We omit the proof. In words, for large n, the distribution of the sum Zn is approximately the normal distribution with mean na and variance nb^2. As with the weak law and strong law of large numbers, this theorem also holds for any distribution for the Yn provided that the variance b^2 is finite. The Standard normal distribution ç is defined by
The corresponding density function / (defined by f (y) = drawn in Figure 3.4.2. Since f (y) = f (—y) for ali y, we have
~y'n^ ) is
(4.16) (— X) = l — <p(x), — <^ X^ <oo.
Figure 3.4.2 The standard normal density function /. In the table of Figure 3.4.3, the value tp(x) is the área under/to the left of x.
Sec. 4
In view of this only.
x
SUMS OF INDEPENDENT RANDOM
relation, it is sufficient to This is done in the table in Figure
<p(x)
x
<p(x)
list the 3.4.3.
x
VARIABLES
values
Ç>«
of <p(x) for x
x
positive
Figure 3.4.3 Values of the standard normal distribution ç for positive x. For negative x use this table along with (4.16).
(4.17) EXAMPLE. Consider a Bernoulli process with success probability /; = 0.5. Then applying (4.14) to the process [Nn\ K e N) of success counts, we get
Next, applying (4.14) to the process {Tn; n & N} of the times of successes, noting'that now a = E(Tn+í — Tn) = l/p = 2 and b^2 = Var(Tn+1 — Tn) = <t/p^2 = 2 by (3.11) and (3. 13),
l • or n = 50, for example, rr ^
is approximately equal to (using the table in Figure 3.4.3)
f^1 -^2 l -/^e~x*;í dx^ = 0.8849.
Similarly,
< 77} = 2.
66 BERNOULLI PROCESSES Chap. 3
is approximately equal to
f"" l -x' f ~ l J _«, A/271 J 2.3 -N/27Z
= 1- f" L..;—
= l - 0.9893 = 0.0107. (^) n
5. Exercises
(5.1) Consider a possible realization a> = (S, F, F, F, S, S, F, S,.. .) of a sequence of independent trials with two possible outcomes, S and F. For this particular ro, what are the values of the random variables (a) XltXt, ...,XS; (b) JVo, #„...,#,; (c) Tt.Ti.Ti.T^TJ (5.2) In Example (1.2) let the defective rate be p = 0.05. What is the probability that the first, second, and third items inspected are ali defective? What is the probability that exactly one of the first, second, or third items is defective? (5.3) For the process described in Example (1.3), compute and interpret the following quantities. (a) P{Ni =0,N 2 = O, N 3 = 1,N 4 = 1}
(c) P{NS=6,N 15 = 12}. (5.4) For the problem introduced in Example (1.4), what is the expected number of bearings, among the first 400 produced, which do not meet the specifications?
(5.5) For p = 0.8, compute (a) EÍN 3 ], £[JV 7 ], E[N 3 + 47V 7 ] (b) Var(AT 3 ), Var(7V 7 - N 3 ) (c) E[6N 4 + N 1 \N 2 ].
(5.6) Consider Bernoulli trials with probability p = 0.8 for success. Suppose that the first five trials resulted in, respectively, S, F, F, S, S. What is the expected value of N 3 + 2N-, given this past history?
(5.7) Generalize the result of Example (2.16) by showing that for any n, m e N, E[N,+m\N»\ = N» + mp.
(5.8) Repeat the steps of the proof of Theorem (2.14) to show the truth of the particular result that
E[3N 5 + NI | N 0 , Nlt N 2 ] = E[1NÍ + NI \ N 2 ]. (5.9) /lypcrtfconwtric distribution. Let k, m, n be integers with k < m -- n. Show l l i. i l l'or ;iny inlcgcr j with O <i j < m, O < k — j < n,
Sec. 5 EXERCISES 67
' y) (m + n
( k ) This defines a probability distribution in j; it is called the hypergeometric distribu- tion. Note that Nm can be replaced by the sum -Xni + • • • + Xnm of any m variables selected from Xlt... , Xm+n. (5.10) In Example (1.2), suppose the items are packaged in boxes of 100 each. A sampling inspection plan calls for rejecting a box if a sample of 5 drawn from that box contained one or more defectives. What is the probability that a box con- taining exactly 4 defectives is rejected? (5.11) In Example (3.1) suppose the rate of crossings is 4 vehicles per minute. Compute and interpret the following quantities. (a) p = P{Xn = 1}, (b) P{T 4 (c)
(d) Var(r 2 + 5T 3 ).
(5.12) Repeat the proof of Theorem (3.6) to show that
if the event [T 1 < 16} occurs. (5.13) Show, by following the steps of Example (3.17), that
(5.14) Compute the following (and compare with (2.2.14)): (a) P{7\ =k,T 2 = m, T 3 = n] (b) P{T 3 =n\Tí=k,Tz = m} (c) E[T 3 \T 1 =k,T 2 =m] (d) £fe(r 3 )| r,, ra] for g(b) = a», a e [0, 1]. (5.15) If a random variable T hás the geometric distribution, then
for ali n and m. Show that the converse is also truc: if a discreto random variable T is such that P[T>n+m\T> n] = P{T > m} for ali m, n e N, then T hás a geometric distribution. (5.16) The probability that a given driver stops to pick up a hitchhiker is p = 0.04; diíferent drivers, of course, make their decisions to stop or not independently ol' each other. Given that our hitchhiker counted 30 cars passing her without stopping, what is the probability that she will be picked up by the 37th car or before? (5.17) Continuation. Suppose the arrivals of cars themselves are as described in l-xample (3.1) with 4 vehicles per minute. Then "success" for the hitchhiker occurs