This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
t 2 )
f
= eit1x g(x) dx
J
eit2Y p(Ylx) dy
296
EXERCISES I N PROBABILITY AN D STA T1STlCS
so that
[or ¢(t~, t 2)]
at
2
= ;r
fe
il "
g(X) Jl~(X) dx.
12 =0
Hence the stated result by using the Inversion Theorem on the fUnction g(x) Jl~(x). Wicksell, S. D. (1933), B, 25, 121. Since
22
1 2(l_p2) [x 2{1-(I- p2)itd + y2{1-(1-p2)it 2} -2pxy]
= _{l-(I- p 2)it d 2(1_p2) therefore
[xpy ]2_y2{(1-itl)(l-it2)-p2(itd(il2)} . 1-(1_p2)it. 2{1-(l_p2)it.} -,
E[exp(it lX2 /2 + it2y2/2)]
= [(1- it d(1- it2)- p2(it l)(it 2)]-i,
whence the joint characteristic function of z. and Z2' Use of the result of the preceding exercise gives 11~(Z d by noting that for N >0
and
(it dO e - il,%, dt 1 (1- it .)N/2 23
Wicksell, S. D. (1933), B, 25, 121. By the preceding evaluation of the joint characteristic function, that of n+n2
'I+nl
w. =
L
xf/20"f
and
j= •
W2
=
L
YJ/20"~
is
j= •
¢(t., t 2) = [(1-itd(l-it 2)-p2(it.)(it 2)]-n/2. (1-it.)-n,/2(1_it 2)-n,fl. Hence by a logarithmic expansion the bivariate cumulants and corr(w., \\'2)
Krs
are obtained.
= KII/(K02 K 20)i.
The conditional moments 11~(W.) are evaluated in the same way as in Ex. 22 above. 24
Wicksell, S. D. (1933), B, 25, 121. To evaluate E[exp(it.x 2+it 2y)], set u _
I 2 2(1- p)
= x/O". and v = Y/0"2' Then
[(~)2+(L)2_2P(X)(L)]+it.X2+it2Y 0". 0"2 0". 0" 2 2(
1 2 [A(u-pv/A)2+B(v-C/B)2_c 2/B], I-p )
where A == 1-2(1- p2)it.O"I; B == 1- p2/A; C == (1- p2)it 20"2'
A !':SWERS AND HINTS ON SOLUTIONS: CHAPTER
4
297
Hence
and so the joint characteristic function of WI and W2 is 4>(t l ,t 2) = (l-itd- n/ 2
exp[Wt2)2{1+;~~;JJ
.
00
= (l-it l )-n/2 ei
L (!p2),(it 2)2r(itd(1-it
(i 12J2
l
)-r/r!
r=O
Ajoint inversion now gives the distribution of WI and W 2 . The variances are obtained by the differentiation of 4>(t l , t 2 ) as in Ex. 22.
Zs Put z), 2 • 1:.+ 111), ) IS
=
~
X), -m), for (1
A.
~ n).
Then the characteristic function of
(I - 2it)-t exp[mi{it)(l- 2itj- I],
dnd that of Y is 4>(t) = (1-2it)-n/2 exp[2112(it)(I-2iW 1 J =
e _,,2 (1- 2iW n/2 exp[Jl2(1- 2iW I J
=
e-,,2
2r
00
L ;(1_2it)-
r=O
whence, on inversion, the distribution of Y. A logarithmic expansion of 4>(t) gives
E(Y)
=
n+2Jl 2 ;
var(Y)
=
211 + 8Jl2.
26 Bartlett, M. S. (1938), JLMS, 13, 62. Let f(x, y) be the joint density function of X and Y, so that
=
f(x, y)
g(xly) h(y)
in terms of the conditional density of X and the marginal density of Y. Then, x,
f
ljJ(t Ily) =
e it1x g(xIY) dx,
-00
and 00
4>(t" t 2)
=
f
ljJ(tlly) h(y) e- it2Y dy.
-00
Therefore, by the Inversion Theorem,
f 00
ljJ(tlIY) h(y)
=
21n:-
e- il2Y 4>(t l , t 2) dt2
-oc,
Jnd
f 00
h(y)
=
2~
-00
,hence the result.
e- it2 ), 4>(0, t 2 ) dt 2 ,
298
EXERCISES IN PROBABILITY AND STATISTICS
27 Kenney, J. F. (1939), AMS, 10, 70.
Since E(e ilX ) = cp(t), therefore the joint characteristic function of Y a Z is nd I/I(t1>t 2 ) = E(eillY+ilZZ) = [CP(tl +t2)]V[cp(td]nl-V[cp(t2)]nz-v.
Further, if h(z) is the marginal density function of Z, then by using the Inv . Theorem, er· SlOn
f e-ilZ'[~. ~I/I] 00
E(Ylz) h(z)
1 = -2 1t
I
ut 1
11:
0
dt 2,
-00
and
.,
E(y2Iz) h(z)
J
= 21 1t
e- i1z '
-00
[~. ~2fJ I uti
dt 2· 11=0
These equations lead to E(Ylz) and var(Ylz), noting that E(X') = [;,.
o'CP~d] ott
I
1,=0
and
f i1z' 00
[cp(t 2)]n
=
e
h(z) dz.
-00
The correlation between Y and Z is obtained from the linear regression equations expressing E(Ylz) and E(Zly). 28
The joint characteristic function of u and v is CP(tl, t 2)
= E[exp(it 1 u+it 2v)] = (1-2it 1 )-nI 2 exp[ -At~/2(1-2itl)], n
where
A
=I
ai.
'<=1
Hence, using Ex. 26 above, the conditional characteristic function or given v = 0 is I/I(tllv = 8) = (1-2it 1)-(n-lI/2. =
ell-
exp[~: {1-(1-2it d}]
f (-(j2Y(1_2it r!
l r(n+2,-l l1 2,
,=0 whence the result on inversion. 29
Bartlett, M. S. (1938), JLMS, 13,62. E(e ilX )
= (1 + t 2) -1.
Therefore the joint characteristic function of u and v is
CP(tl, t 2)
= E[exp(it l u+it2V)] = E[exp i(tl +t2 )xtl. E[exp i(t l -t 2 )X2] =
[1+(tt+t 2)2]-l[I+(tl-t 2)2r 1.
Hence evaluate cp(t dv) by Ex. 26 above.
1/
ANSWERS AND HINTS ON SOLUTIONS: CHAPTER
4
299
for v > 0, evaluate the integral 00
f
eit2V cf>(t 1, t 2) dt2
-00
by using a semicircular contour in the upper half of the complex plane.
fherelevant poles are at t2 = i±t1. Hence the numerator in cf>(t1Iv) is ne - v [t 1 cos vt 1 + sin vt 1 ] t 1(1 +tn
2'
.
The denominator is ne- v(1+r)/2. obtained as the limit for t 1 --+O of the numerator. 30 Bartlett, M. S. (1938), JLMS, 13,62. Use the result of Ex. 26 above.
f
00
Je- it2Y
00
cf>(t1' t 2) dt2
=
-00
J
exp{ -it 2Y+K(t 1, t2)}' dt2
-00
00
eoP[it,,-(oloy)). exp[ - it 2y -!(K20t i + 2K 11t 1t2 + K02t~)] . dt2
-00 00
=
f
eoP[it!. -(016y)) . e-tK2ott
exp['t 1( K02 t 2 - I 2Y -2 2 + 2K11 t 1 t 2 )] • dt 2
-00
whence cf>(t 1IY)· 31 The characteristic function of the sample mean
[eit;;/:
x is
I]",
and so by the Inversion Theorem the probability density function of x is f(x)
= -
1
Joo
2n
itl e- itx [e- ."-1till
1]" dt.
-00
The integrand is analytic everywhere. and so the path of integration may be changed to the contour r consisting of the real axis from - 00 to - p, the small semicircle of radius p and centre origin, and the real axis from p to 00. Thus itl f(x) = -1 e- itx - .dt 2n 1till
f
[e "_I]"
r
=
.(11) . feit[(jf,,'-X) dt.
(-1)". " (-1)) j
2n
j~O
(i/Il)"
r
t"
300
EXERCISES IN PROBABILITY AND STATISTICS
But ei~Z
f -Z"
dz
= 0, for 0( > 0,
r
2nO("-1 - i" . (II _ 1)!,
for 0( <
o.
Hence the value of I(x). 32 00
E(e itX )
= 2~
f
f 00
eitx-Ixl/a dx
=~
e -xla . cos xt dx
0
-00
= 1-(12t 2 . E(e itX ). Hence the characteristic function of x is (1 + (12 t2/11 2)-II. Therefore the prob. bility density function of x is ,t·
-00
-00
The integral is evaluated by using a semicircular contour with a pole Ilf order II at t = in/(1, whence the distribution of x. Direct integration gives
f 00
g(x) dx
=
-00
II - 1 (II +,. - 1) 1 L -1 2"+r-1 r=O II II-I (1 +O()"+r-I r~o -2-
= coefficient of 0("- I in
1+0()"-1 211-1 (211-1) ] 2 [(- O(r/2 211 - 1 2 r=O 1
L
= coefficient of 0("-1 in
1 = 2 [1- .,211I· -
LI
II -
r=O
(211 - 1)]
.
.
L rl JJ
5=0
' whence the result.
I
33
Kullback, S. (1934), AMS, 5, 264. The characteristic function of u is aitllO + it)-II, and the probability density function of u is
f
~
h
_ 1 I(u) - 2n
exp( - itu + itll log a) dt __1_ (l + it)" - 2ni"
f
eitA dt , (t - i)"
-00
-00
where A. == II log a - u ~ O. The integral is evaluated by using a semicircular contour. whence The distributions of g and 2n log(a/g) follow by transformation. 34
E(eitIXI) =
(l-(1it)- I, so that the characteristic function of v is (1-(1it/n)-1I
whence the result.
= {1-2it(:n)} -n,
ANSWERS AND HINTS ON SOLUTIONS: CHAPTER
~ ECeiIX) I. . Hence
=
(1
4
301
+ (2)- I.
[exp i(t I U + 12 V)] = E[exp i{(t 1+ t 2)X + (t 1- t 2)Y}] = [I +(tl +t 2)2r 1[1 +(tI-t 2)2r I. rherefore the characteristic function of u and (Wl>t 2 )
= [1+(tl:t2rrn
v is
[1+el:~2rrn.
rhUS the characteristic functions (W 1> 0) and cP(O, t 2) of u and v are of the ..Jdle form, but cP(t I> t 2 ) 1= cP(t I> 0) cP(O, t 2 )· Expansion of 10gcP(tl> t 2 ) shows that the coefficient oftlt2 is zero. The probability density function of u is
-00
-00
rhe integral is evaluated by using a semicircular contour. Hence _
ne- n1iil
2n-1 (2n+r-1)!
g(lI) = (2 II _1)1' .
L
r = ()
-1(" _'_1)1' , • _II 1 .
[nluWn-r-1 _ ,)2n+r • (-r:o
J6 Kullback, S. (1934), AMS, 5, 264. E(eitIOgX)
= r(p+it)jr(p).
Therefore the probability density function of u is 00
g(u)
=
-1 2n
f e -,'U. [rIp + it)/r(p)]n dt
-00
ePU f-p+ioo = 2ni[np)]n' eUz [n _·z)]n dz. -p-i'X)
But nz)r(1-:) = nisin nz, so that
f
- p+ ioo
e PU( -nl" 1 g(u) = .[r(p)]n 2ni
eUz dz .' {n1 + z)}" smn nz
-p~ioo
Evaluate the iptegral using the contour Iz/ = m+t, m a positive integer, and the line x = - p. The poles of the integrand are at z = r with residue
[d
n- I ] n-n( _1)nr eUZ (n -1) ! dzn- I . {n1 + z)}"
z=r •
Hence
ePU 00 (_1)nr+n+ 1 [d n- I eUz ] I n g(u) = [r(p)]n' r~o (n-1)! dz - ' {n1 +zW z=; Whence the distribution of v.
302 37
EXERCISES I N PROBABILITY AND ST A TISTI CS
Let
'IY r(Z+~)I/lIt-"Z(21t)(n-Il/2nIlZ). V
R(z) =
,=0
II
Then R(z) = R(z + I) so that R(z) = !~~) R(z). But for large z, Stirling's formula gives r(z+1) = (21tZ)t(;r+ O(Z-I). Hence log R(z + 1) = O(Z-I) so that lim log R(z+ 1) = O. • -+ 00
The duplication formula is obtained for 38
II =
2.
Kullback, S. (1934), AMS, 5, 264. Since E(eil)ogXj) = nPj+itvnPj), the probability density function of 1/ is
•
i 'X
feu. Ii r(pj-Z)' dz.
n 1 .~ r(pj) 2m -ioo
fl
j=1
j= 1
The integral is evaluated by using the contour Izl = m+t, where m is an integer. and the imaginary axis. For fixed j, the poles of r(pj-z) are at z = Pj+r ror r = 0,1,2, .... The residue at z = Pj+r is (-1)'+ 1 e U('+ Pj) nr+ 1) . IJj r(Pk-pj-r).
1
Hence g(u)
=
n
00
·.L L fl r(pj) )= 1 ,=0
(_1)'+2 eu(r+Pj) r('+ 1)
n
I
.
fl. r(Pk- pj-r), k'¢)
j= 1
whence the distribution of v = eu/n• 39
Kullback, S. (1934), AMS, 5,264. Since Pj = p+(j-I)/n, by Ex. 37 above n
fl
r(pj) = nt -np(21t)(n-ll/2nllp).
j= 1
Therefore the characteristic function of u is n -nilnnp + nit)/nnp),
and the probability density function of u is
()- f
~ g u - 21t
00
e
-ilu n-nilnllp+nit) d . nnp) . t
-00
-np+ioo
=
1 - (-)'-2' nr np m eUPn"P
f (n eu/nyn-z)dz.
-np-ir
ANSWERS AND HINTS ON SOLUTIONS: CHAPTER
4
303
fb e integral is of the type -a+ ico
I ==
~. Jwz f'( -z) dz. 2m -a-ioo
"/0 and w real, which is evaluated by using the contour Izl = m+t, m a ~sitive integer, and the line x = -a. The residue at the pole z = r is 1~\V)'/1tnr+ 1), so that I = e- W • Hence the distribution of v = e"/n is 1 - - . e-m'(nv)nP-'n du,
(0 ~ v < 00).
r(np)
If
Xl'
X2" .. , Xn are random observations from a n
parameter p, then
I
has the
Xj
r
r
distribution with
distribution with parameter np. Hence the
j= ,
distribution of the sample mean i. ~O
Fisher, R. A. (1929), PLMS, 30,199, and UL (1964). .
E(e" x )
(it)'
7J
=
1+
I 2 Ilr'-" r.
r=
"I
that the cumulant-generating function of i is
I [1 + ~ Ilrn
n og
L...
r'
r=2
(it)'] ,
r.
.
By expansion
-) = -11-2 "2 (X
n
an d
Again, n
(n-l)s2 =
I
xJ-ni 2, so that E[(n-l}s2] = (n-1)11-2'
j= 1
Also, n
(n - 1)2s4 -_ "L... X4j + j= 1
"L...
X2X2 j k-
j*k
so that _ (n-l)2 (n-l)(n2-2n+3) 2 E[(n- 1)24] s - ---·11-4+ ·11-2'
n
n
41 Lukacs, E. (1942), AMS, 13,91. If cf>, (t ,) is the characteristic function of i, then 4> 1 (t d == 4>(t h 0) == [I/!(t dn)]n ;
and. hy definition. 4> 2(t 2) == 4>(0. t 2).
whence var(s2).
304
EXERCISES IN PROBABILITY AND STATISTICS
The independence of x and S2 means that (p(t I' t 2) == ¢I(t .)¢2(t 2) Or
But
a¢ = I.
-? t2
f
S2
. S .exp (.1t 1X+1t 2 2). TIn f( X,. ) dXv>
,.= 1
f(x) being the density function of X. Therefore
[ata¢]
= i["'(ttfn))"-1 feiIOC/n.x2f(X)dX-
2 12=0
_;["'(t.ln)]n-2 [feiIIXln.Xf(X)dxr Hence the differential equation by putting ttfn = t. The solution
follows since ",(0) = 1 and ""(0) = ill.
42 Williams, J. D. (1941), AMS, 12, 239, and McGregor, J. R. (1960), B,47, 111. The variables Yj = (xj-m)/a are unit normal, and n-l
A
=!
L (Yj-Yj+l)2; j= 1
n
B
=!
L(yj-ji)2. j= 1
Therefore
But n
-!
L YJ+tlA+t2B j= 1
where a == 1-tl -[(n-1)/n]t2; b == tl -t2/n; c == 1-2tl -[(n-1}/n]t 2; d == /2/11 so that a+b+(n-2)d = 1: 2b+c+(n-3)d = 1. M n has the leading diagonal (a, c, c, . .. c, a), bordered by diagonals whose c\.cments are all h's, and the remaining elements are d's. Row operations gIve Mn = t~-1 Lln' where Lln is the nth-order determinant:
ANSWERS AND HINTS ON SOLUTIONS: CHAPTER
1 1
1
1 1 1
()
1 0
0 0 0
1
()
1
0 0 0
0 0
1
()
0 0 0
0 0 0
0
1
0
.......
0 0 0 0
()
4
305
1
1 1 (0+ 1) ,
.here 8 == (c-d)/(b-d); ()+ 1 == (a-d)/(b-d). Let D.-l denote the (n -1)th-order determinant obtained from An by xcluding the first row and column of An. Also, let C j denote the jth-order ~elerminant formed by the first j rows and columns of Dn _ 1, U < n - 1). Then Mn=t~-l(Dn_l-An_l)' C 1 =(}, C 2 =(}2-1, and for j>2 II
(.::: (}C j - 1 -C j J
2•
Hence
BUI
and /).n = Dn- 1 -An Therefore
1•
Assuming that An = AIZn+A2Z-n, the constants Al = - z/(1- Z2); A2 = z/(1- Z2). Hence An = An- 2 +S n- 1 , where S. == z'+z-', whence
Mn = t~Mn_2+~-lSn_l' Bul z, Z -
1
are the roots of Z2 - (}z + 1 = O. Therefore Sn-(}Sn-l +Sn-2 = 0,
and the difference equation for Mn follows. j
To obtain the expression for mj' put t2 =
f 00
= [ 0i>(t 1;t 2)] ot 1 1,=0
A j e/2B •
-00
L
,= 1
t,
in
fIk=l -1-.e-tY~dYk' fo
and integrate for ( - ex) < t. ~ 0). Also, for the given form of >(t 1> t 2)
[ oj>(t~,t2)] ot 1
= [dj>(tt O)] I, =0
dtl
. (1_t 2 )-tln + 2 j-l). I,
=0
306
EXERCISES [I>< PROBABILITY A;,\;[) STATISTIC'S
43 Cochran, W. G. (1937), JRSS, 100,69. Let uj = xj-x and Vj = Yj- y. Then E(uj) = E(vj) = 0; var(u) = var(vj ) = (n -1)/n; and corr(u j , Vj) = p.
Also, since uj and Vj have a joint bivariate normal distribution, P(u j > 0, Vj > 0)
=
P(uj < 0, Vj < 0)
=
!+(1/2n) sin-I p
by Sheppard's medial dichotomy theorem. Hence E(C) = t+(1/n)sin- 1 p and var(Cj ) = !-[(1/n) sin-l p]2. Note that, by symmetry, P(1I 1
> 0,
VI
> 0; U2 > 0, V2 > 0) = P(UI < 0,
VI
< 0; U2 < 0, v2 < 0).
To obtain P(UI > 0, VI > 0; U 2 < 0, V2 < 0), change the sign of t3 and I . the characteristic function. Then inversion and integration now give 4 III P(u l > 0,
VI
> 0; U2 < 0, V2 < 0)
1
1
= 4n 2[(sin- 1 p)2-{sin-1 p/(n-1W]+ 4n[sin- 1 p-sin- l p/(n-l)]+conslalll = P(u l < 0, VI < 0; u2 > 0, V2 > 0), by symmetry.
Hence cov(C I, C 2 ) = constant - [(lIn) sin -I p/(n -1)F whence var(C), the constant being evaluated since var(C) = for p = 1. For large n,
°
var( C) - :n [1 -
(~ sin - I P ) 2].
which gives the large-sample efficiency of p*. 44 Bartlett, M. S. (1936), PRS, 154, 124. E[eit(X-m)]
= (1+t 2)-I, and y-m = (1-JL)(xl-m)+JL(x 2-m).
Therefore the characteristic function of (y-m) is E[ exp{ it(1- f/)(x 1- m)}] . E[exp{ itJL(X2 - m)}]
= [1 + t 2(1- JL)2] - I [1 + t 2p2]- I 1
= 1-2JL
[
(1- JL)2 1+t2(1-JL)2
JL2] 1+t2JL2'
But (1 + t 2 a 2 )-1 is the characteristic function of the distribution 1 - e- 1zl /C1 dz 2a' ,
( - 00
<
Z
< (0).
Hence the distribution of (y - m) is 1
2(1- 2JL)
[(1- JL) exp{ -ly-ml/(l- JL)} - JL exp{ -ly-mI/JL}] d(y-m). ( - 00
< y < (0).
Al"SWERS AND HINTS ON SOLUTIONS: CHAPTER
4
307
fherefore the logarithm of the likelihood is log L = constant + log[(l- JI) exp{ -ly- ml/(1- JI)} - JI exp{ -1.1'-1111//1]]. ,0
that
1
[1
L
] ,,_ . -+y~ 2 +(y_l)2 00 y' foo exp-[{(r+2){3-(r+l)lX}z].dz, (1- 2JL)(I- JL) IX P r= 0 o
IP> IX). whence the result.
SUPPLEMENT TO THEORETICAL EXERCISES IN PROBABILITY AND STATISTICS
Supplement Finite differences and summation of series Sum the following series by the method of differences: (i)
t
L
r(r + 4)(s - l)(s + 2)
r= 1 s
(ii)
L" L r(r + 3)s(s -1) r=1 s=S;;r n-l
(iii)
L L r(r+2)(r+4)s(s-1)(s+2)
r= 1 s=S;;r
(iv)
t
L (r+1)(r+2)(r+4)s(s-1)(s-3)
r= 1 s=5;;r n+l
(v)
L L (r+2)(r+6)(s-2)(s-4).
r=1
s
Sum the following series by suitable methods: (i)
L" (r -
1)2(r - 2)2(r + 3)
r=1
(ij)
L" r. 2'/(r + l)(r + 2) r= I
(iii)
L'" (r 2+3r-6)/r! r=l
(iv)
L'" (r3+S r2+6)x
r•
r=O
Sum the following finite binomial series: (i)
'f (r+n 1)r(r + 2)x
r
r=O
(")
" r r(r -l)(r + 3)x r • (ii) r~ Given that the rth term of the series
3,4,7,14,27,48,79, ... is a polynomial in r, find the general expression for the rth term. Hence derive a particular value for this term and sum the series to It terms by using Ihis value. 311
312 5
EXERCISES IN PROBABILITY AND STATISTICS
By using Euler's method of summation, sum the following series' 9 25 9 r2 (i) 1+2+-+2+-+-+" .+-+... 4 16 8 2,-1 '
and
00 1 27 125 27 r3 (11) -+1+-+2+-+-+" .+-+ ....
4
16
64
16
6. For a series of terms Uh Wise, that (i)
U I1 +1
= (1 + A)"Uh
(ii) AlI1 u I =
2'+1
U2, U3, ••• , UI1 ,
•••
prove by induction, Or other. .
and
I (-1y(I1l)U r
,=0
II1
+ 1 _"
~her.e, in standard difference notation, A111 +1 U, = AII1 (U,+1 - u,), for all posi.
tIve mtegral values of 111. Also, by considering the differences of the series
where U,
= A-1u,+1 - A-1u"
show that 11 U,=A-l[(1+A)I1_1]Ul='~1 11 (") (iii) '~1 r A'-IU I · 7
A function f(x) is tabulated for a sequence of values x, x + h, x + 2h, ... , x + rh,
If it is known that A2f(x) = 0 + {3x, where
(r;;;. 3).
and (3 are constants, prove that
0
f(x) = ao+ alx + a2x2+ a3 x3
where ao and al are arbitrary constants, but a2 = (0 - {3h)/2h 2 and
a3 = (3/6h 2.
8 For a function f(x) tabulated at the points x, x + h, x + 2h, ... , define the finite difference operation A'f(x) for positive integral values of r. Hence prove that
(.) A2tan-1 x=tan-1[
-2h 2(x+h) ] {(x+hf+1}2_h 2{(x+h)2_1}
I
(ii) [A2 +2(1-cosh)(1+A)]sinx=O, and ( 00') [All -1 III A- 1
]A/3
X
= [(A/3h -1)11 -1 A/3 h - 2
where A and {3 are constants and 9
It
]A/3
X
'
is a positive integer.
If A is the standard finite difference operator with the argument interval
SUPPLEMENT
:: (7 0)
313
operating on the function f(x), define arf(x),
for
r?3l,
lIenee prove that, for h sufficiently small so that terms involving h 3 are ·legligible, (i) a2 10g(1+x)=-h 2 /(1+xf (ii) [a2 + h 2 (1 + a)] cos x = 0, and (iii) a2ex2 = ex2 • 2h 2(1 + 2x 2). (0 (i)
If u" the rth term of a series, is such that Ur == fer) - fer + 1), prove that the sum of the first 11 terms of the series is
f(l) - f(1l + 1).
Use this result to find the sum to 11 terms of the series for which 1 (r ?31). (3r - 2)(3r + 1)(3r + 4) Hence deduce that the sum to infinity of the series is
214.
(ii) Suppose Ur == arx r are such that
a r + aa r_\ + {3a r-2 = 0,
for r?3 2
and a, {3 are constants. If S denotes the sum of the series uo, Ul> U2, ... , then by considering the product (1 + ax + {3x2)S prove that S=ao+(a\+aao)x 1 +ax+ (3x 2 Hence, as a particular case, determine the sum to infinity of the series 1-8x +28x 2-80x 3+ . .. , and indicate the range of x for which the summation is valid.
°
11 (i) For a> and Il small such that terms involving show that a root of the equation is a (1-!1l log ex). (ij) Prove that for any r>
11 2
are negligible,
°
1 r(r+2)
2r+l 2r+3 2r(r+ 1) 2(r+ 1)(r+2) .
By using this identity, or otherwise, find the sum to Il terms of a series whose rth term is 1/r(r+2). What is the limiting sum as Il ~ oo? (iii) For any two numbers a and b such that a > b > and a - b is small compared with a, prove that the value of the function
°
g(a, b) = a;::2
-log(~)
lies between (a-b)3/6a 3 and (a-b)3(3a-2b)/6a 3b.
314
EXERCISES IN PROBABILITY AND STATISTICS
12 (i) If f(r)
= X-I /
(x: r) where
is any positive integer, prove that
X
f(r-1)-f(r)=r- 1 /(x:r),
for r;;;'1.
Hence show that
so that, as n ~ 00, the sum of the series tends to X-I. (ii) If X is a sufficiently large number such that terms involving · 'ble, prove t h at negI19l (
(iii) If
Ixl < 1 and
X
x-4 , tlrt'
-I)! = 1-~+ 1 1 1 2x 2 - 2x3'
x+1
the rth term of an infinite series is
(-xY Ur =
prove that
r(r+ 1)
for r;;;.l,
L u =l- (1_+_x) 10g(1+x). 00
r
X
r=1
13 (i) If the rth term (r ;;;'1) of an infinite series is 1 +2+2 2 +, .. +2 r - 1 U = r (r+ I)! prove that 00
LU
r
=!(e-1f.
r=1
(ii) Prove that for any positive integer
"f (r-1n ) (3r
r
r=1
)
n;;;'
=
1
4"+1-1. n+1
(iii) If x is sufficiently large for terms involving x- 4 to be negligihk. prove that 2 1 1 (x + X)2 [Iog(x + 1) -log x] = 1- 24x2 + 24x 3 . 1
(iv) If x is sufficiently small for terms involving that
XS
to be negligible, prow
14 (i) The function f(r) = 1/(1 + r+ r2) is defined for all non-negative val· ues of r. Prove that f(r-1)-f(r)=1
2r 2 4' +r +r
forr;;;'1.
315
SUPPLEMENT
Hence show that for any positive integer S ==
i r=
"
2,. 1
11
11(11
1 + ,.2 + 1'4
+ 1)
n 2 + 11 + 1 '
so that lim S"
=
1.
.. +--GO
(ii) If
Ixl
~1
=
X
-
1 (1- x)( 1 - 2x) .
(iii) If 0 is a small number such that 03 and higher powers of 0 are negligible, prove that an approximate value of the positive root of the equation eX +x = 1+0
is
x = to(I-iO).
IS (i) If the rth (1';;;0 0) term of an infinite series is Ur =
41'3+211'2+30,.+ 11 (r+3)!
prove that '"
L U = 4(e-l). r
r=O
Hence show that if a discrete random variable X has the probability distribution with point-probabilities proportional to u" then
P(X;;;o2) = 48e-l03. 48(e-l) (ii) If a and (3 (a;;;o (3) are any two positive integers, prove that
at k=O
(a-k)= (a+l). (3 (3 + 1
(iii) If n is a positive integer and
Ixl < 1, prove that the coefficient of xr (n+~-l).
in the expansion of (1-x)-" is
t k=O
k
(a + -1)((3 +a k
a
Hence show that
=k -1) (2a + (3-1), =
k
a
where a and (3 are any two given positive integers.
16 (i) If a is a positive constant and
"
r~o (-1)'
11
a positive integer, prove that
(n) (1+ra)/(1+na)'=O. I'
(ii) If u" the rth term (1';;;01) of a series, is
16
316
EXERCISES IN PROBABILITY AND STATISTICS
evaluate the sum of the first /l terms of the series, and hence . that the sum to infinity is unity venfy (iii) Given that f(r)==or 2(,.-1)2(2r-1)(r 2 -r-1), prove that r6 = tr2 + 1[f(r + 1) - fer)].
Hence show that for any non-negative integer r II
L r6 = /l(/1 + 1)(211 + 1)(3/1 4 + 6/1 3 -
311 + 1)/42.
r=1
17 (i) For any two positive integers 11 and s (s < /1), prove that L (r+s)(s+J) = -1- (11 +s+ 1)(s+2). r=\ s+2 II
Hence, if Sr denotes the sum of the first r natural numbers, that
ShOll'
II
L SrSII-r = l~o(11 + 3)(5), r=\
in standard factorial notation, 11 (;;.1) being an integer. (ii) Prove that for any non-negative integer 11 II (11)/ L (r+2)= 11 . 211+1 + 1 . r=O r (11 + 1)(11 + 2)
18 Prove that (i) the average of the first 11 (;;.1) natural numbers is t(11 + 1); and II
(ii)
L [r-!(11+1)]2=-b11(112-1). r=1
(iii) If r is a discrete random variable which takes the values 1,2, ... , /I with probabilities 11- 1 , then by using the exact expressions for E(Jrl and var(..r;.), prove that [1- 2V(I1) ], [ t Jr]2 =11 2(11+1) 2 11(11 + 1) r=1
where "
V(I1) ==0
L (Jr-O)2
11
and
r=1
0=
L 0-/11. r=\
19 Explain clearly the difference method in the summation of finite and infinite series. (i) If U r = 3r4 (r -1)4 - 4r 3 (r -1)3 for all integral values of ,. ;;.1, prO"1.' that II
Lr
7
= -b11 2(11 + 1)2 + {4UII+\
r=\
( oo) If II
Ur =
2r+ 1
r 2( r+1 )2 '
evaluate the sum to 11 terms of the series deduce that its sum to infinity is unity.
U 1 , U2, U 3 , . . . ,
and henel.'
SUPPLEMENT
317
(iii) By using the identity
.L .L. II
XjXj
1 [(
=-2
1=1/<1
.L )2 -.L x~ II
II
Xj
1=1
]
,
1=1
prove that the sum of the products in pairs of the first n odd integers is in(n-1)(3n 2 -n-1). !O (0 Define the finite differences of a function [(x) tabulated for equispaced values of the argument x. If, for x = r, the value of [(x) is [(r) = sin«(J + rq,), where (J and q, are constants, then prove that for any non-negative integer n L1"[(r) = (2 sin(q,/2))" . sin[(J +!n('1T + q,) + r
Uo, Uh U2, .•• , are the successive values of a function corresponding to those of an integral variable r ~ 0, then prove that for any non-negative integer n,
L1"uo=
~ r=O
(n)(-l)"-rUr. r
Hence, or otherwise, show that if ur =(-lY/(r+1), then Llluo=(-1)1(2"+ 1 -1)/(n + 1).
(iii) Find the sum to n terms of the series whose rth term is
ur =(r2 +r-1)/(r+2)!, for r~1.
Also, verify that the sum to infinity of the series is
!.
ProbabDity and statistical theory 11 At a university in Ruritania the probability of a first-year student passing the June examination is t and that of receiving a res it in the following September :&. The probability that a resit student will pass the September examination is~. Prove that the probability of a first-year student ~ntering the second year at the university is ~. If the probability of a second-year student ultimately getting a degree is ;~, determine the initial probability of a student receiving a degree at the end of three years. Hence evaluate the probability that of n randomly selected first-year entrants to the university (i) exactly k students will receive a degree; (ii) at least k students will receive a degree; and (iii) at least k students will not receive a degree, for O::s:; k ::s:; n. 11 At a game of poker, a player is dealt out a hand of five cards from a Itandard pack of 52 cards. Find the probability that the player has (i) a flush, that is, all cards of the same suit; (ii) at least one spade; (iii) at least one spade, but not a flush; and (iv) a flush, but not of spades.
318
EXERCISES IN PROBABILITY AND STATISTICS
Hence determine the probability that the player has (v) a flush when it is known that he has at least one spade' and (vi) a flush of spades when it is known that he has a flush. '
23 An electrician buys a box containing N fuses of which a number '
1)]':
and (b) the conditional probability that a second sample of v fuses from till' N - v remaining ones is wholly good is
r~l
(;)
(~ ~~v) / (~
1_(N:n)/(~ 24 A bag contains 6n (n;:. 1) identical tickets which are numbered serially from 0 to 6n -1; and an experiment consists of drawing a ticket randoml\' from the bag and noting its number, This experiment is repeated three timc~, the drawings on the second and third occasions being made after tl1l" previously drawn ticket had been replaced in the bag, If S denotes the slim of the numbers on the three tickets drawn, prove that d ( ') P(S=6 )=(3n+2)(6n-l) , 1 n 216n 3 ,an
(ii) E(S) =~(6n -1). Also, show that if the sampling is without replacement of the drawn, then 3n P(S = 6n) = (6n -1)(6n -2)
tickct~
25 An urn contains m + n balls which are identical apart from their colour: m of them are black and the rest white. A ball is drawn at random froIll the urn and its colour observed. The ball is then returned to the urn and. in addition, r balls of the colour drawn are added to the urn. A ball is 11011 drawn at random from the urn. Find the probability that (i) both balls drawn are white; (ii) both balls drawn are black; (iii) one ball drawn is black and the other white, irrespective of the ortle' in which they are drawn;
SUPPLEMENT
319 (iv) the second ball draw.n is whit~; ~n? (v) the first ball drawn IS black, If It IS known that the second ball is white. I\lso determine the expectation of the number of white balls drawn in Ihe tWO trials. '6 An urn contains 2N balls which are identical except for their colour, : If the balls being white and the rest black. If r «N) randomly selected
:ans from the urn are all found to be of the same colour, determine the ~obability that the colour of the balls drawn is white. r In another experiment k «N - r) black balls are first removed from the ~rn, and then, from the remaining 2N - k balls, r balls are selected ranJomly. If these selected balls are known to be of the same colour, show that Ihe probability that the colour of the balls drawn is white is 1 p = 1 +(N-rYk)/N(k)'
If N is large compared with rand k, so that the terms involving N- 2 are negligible, prove that approximately p=
~ [ 1 + 2k~ J.
17 In "The Final Problem", a short story by Sir Arthur Conan Doyle, Sherlock Holmes is pursued by his old enemy Professor Moriarty with murder in mind. It is understood that if Moriarty catches Holmes then the latter must die. Holmes takes a train at London bound for Dover from where he can escape to the continent. Moriarty charters a train and follows dose behind. There is only one stop before Dover-Canterbury. If both Holmes and Moriarty choose to alight at the same station, then Holmes meets his end. Secondly, if Holmes chooses Dover and Moriarty Canterbury, then Holmes makes good his escape. Lastly, if Holmes gets off at Canterbury and Moriarty goes to Dover, then the pursuit is a draw since Holmes, though temporarily safe, is still within the reach of Moriarty. If the independent probabilities of Holmes and Moriarty getting off at Canterbury are x and y respectively, determine the probabilities of the three distinct outcomes of the pursuit. If the probability of Holmes meeting his death in the encounter is a (>0), and this is also the probability of his escape, then prove that y2-3ay+a =0. Hence deduce that in this particular case a must satisfy the inequality 28 (i) A standard pack of playing cards consists of four colour suits, each suit having 13 distinct cards. The 13 cards of a suit include four court cards-an ace, a king, a queen and a jack. From the pack, a hand of 13 cards is dealt out randomly to a player. Find Per), the probability that the hand contains exactly r court cards, 0:;:; r:;:; 13. Hence show that, for r ;;'1, per) satisfies the difference equation P( )_(17-r)(14-r)p(
r -
r(23+r)
P(O) = 36(13)/52(13\
).
r-1,
and
in standard factorial notation.
320
EXERCISES IN PROBABILITY AND STATISTICS
(ii) Two players A and B playa game of chance as follows. From the pack of 52 cards, 13 cards are dealt out to A and then another 13 cards to B, the player having more court cards being the Winner f the game. The game ends in a draw if both A and B have no COU~t cards or the same number of them. Prove that the probability fo the game to be drawn is r
t (16)(16-k)( 36 )(23+k)/(52)(39) k k 13-k 13-k 13 13'
k=O
Also, in particular, if it is known that A has all court cards, then show that the probability of B having no court cards is ~ approximately. Verify that the percentage error in this approximation is;'!7·
29 In a game of chance, a player repeatedly rolls k (> 1) unbiased cubical dice till at least one six shows up. If the event "at least one six" OCcurs on the rth throw of the dice, the player wins £2 1- kr (r;;;:.l). Prove that the probability of the player winning on the rth throw is (~)k(r-l)[l-(~)kJ.
Hence deduce that his expected reward is l-(~)k
£ 2k -
1
[1-(&YJ'
so that, on the average, the player can never be on the winning side if the non-returnable stake to play the game is £2 1- k • Also, show that the expected reward is a decreasing function of k.
30 An urn contains n (> 3) balls, either white or black, but otherwise identical; each ball has a probability p of being white and q of being black (p + q = 1). A random sample of k «n) balls is drawn from the urn. Prove that the probability of the sample containing only one white ball is k
"-t+ (nr-11
k)prq"-r.
r=1
Furthermore, if it is known that the urn contains one black and one white ball and the remaining n - 2 balls are equally likely to be either white or black, prove that the probability of the sample of k balls containing only one white ball is ,,-k+l (11 -
k
n(I1-1)2,,-2'
r~1
k)
r(n-r) r-1 .
31 A man plays a game of chance in which, at each turn, he rolls a pair of unbiased cubical dice whose faces are numbered from 1 to 6. He continues until the first "successful" turn, a successful turn being one in which at least one of the dice shows 6. If this is his 11th turn (11 ;;;:.1), he receives a reward of £a" (a < 1). This terminates the game and the stake for playing it is 50 pence. Prove that the player's expected reward is £l1a/(36-25a). Hence deduce that, on the average, the player will be on the winning side if a> ~~. 32 In a game of poker, five cards out of a standard pack of 52 are dealt out at random to a player. There are four suits in a pack and the 13 cards in a
SUPPLEMENT
321
suit are regarded to be in an ascending order in which the ace is taken to be Ihe lowest card and also the highest. Any five cards (not necessarily of the same suit) which are in an unbroken sequence are said to form a run. I-Iowever, if the five cards in the run belong to the same suit, then the sequence is known as a running flush. Find the probability that the player has a run but not a running flush. Also, show that there is a 4 per cent increase in this probability if it is known Ihat the player has the jack of hearts. What is the explanation for this increase? 32a There are It alternatives in a multiple choice test, and the probability of an examinee knowing the right answer is p. However, if he knows the right answer, then it is not absolutely certain that he will give the correct response because of the emotional tensions of the testing situation. It may therefore be assumed that the probability of the examinee giving the correct response when he knows the right answer is 1 - a where a > 0 is small but not negligible. Alternatively, if the examinee does not know the right answer Ihen he randomly selects one of the It altet:natives. (i) Prove that the probability of the examinee knowing the right answer when he, in fact, gives the correct response is Itp(1-a)
l+[n(l-a)-l]p·
Also, determine the probability of the examinee not knowing the right answer when, in fact, he gives an incorrect response. (ii) Evaluate these probabilities on the assumption that when the examinee does not know the right answer, he knows that m < It of the alternatives are incorrect so that he makes a random choice from the remaining It - m alternatives. (iii) Find the limiting values of the probabilities in (i) and (ii) as It ~ 00, and comment on the results obtained. 33 As a financial inducement to Ruritanians for changing their habits of smoking, the government presented two separate proposals in parliament. The first proposal referred to a decrease in the excise duty on cigars and lobacco, and the second called for an increase in the duty on cigarettes. Two free votes were taken in parliament with the following results. The ratio of Ihe number of M.P.s voting for a decrease in duty on cigars and tobacco to Ihe number voting against it was one and a half times the ratio in the vote on an increase in the duty on cigarettes. Again, of those M.P.s who voted in favour of a reduction in the duty on cigars and tobacco there was a majority of 165 in favour of the increase of duty on cigarettes; and of those who voted against the decrease in duty on cigars and tobacco there was a majority of 135 in favour of an increase of duty on cigarettes. Finally, if 110 of those who voted for both government proposals had, in fact, joined those who voted against them, the number of the former would have been twice Ihat of the latter. Assuming that there were no abstentions in the votings, find (i) the total number of M.P.s attending the meeting; (ii) the numbers voting for each proposal; (iii) the numbers voting for and against both proposals; and (iv) the numbers voting for one and against the other proposal.
322
EXERCISES IN PROBABILITY AND STATISTICS
34 Each of two numbers x and y is selected randomly and independ from the integers 1, 2, 3, 4 and 5. Show by suitable enumeration tha~ntly probability that x2 + y2 is divisible by 5 is is. Further, prove that this is ~he the probability for 2X2+3y2 to be divisible by 5, and that the probab~l'sll that both x 2+ y2 and 2x 2+ 3y2 are divisible by 5 is ,Js. Illy Extend these results to the case when x and yare chosen frollJ h integers 1 to 5N, where N is any integer ~1, to show that the ab t e probabilities are independent of N and so also applicable when x and y ~\'e selected from the set of all integers. dre
35 There are k similar urns numbered from 1 to k, each containing IJJ + 1 (11 > k) balls which are identical apart from their colour. The balls in the urn are all white, but in the rth urn (1'::;; r'::;; k -1) there are m + r black 'I ; 11 - r white balls. An urn is selected randomly and two balls are drawn fl:(;l( it. If both the balls are found to be white, determine the condition~; probabilities that the kth urn was selected, on the assumption that tI;~ second drawing was made (i) after replacement of the first ball drawn; and (ii) without replacement of the first ball drawn. Explain very briefly why the conditional probability in (i) should be less than that in (ii).
kl:
36 There are six unbiased dice Dl> D 2 , • •• , D 6 , the six faces of D j beinl: numbered (i, i, i, 4,5,6) or (i, i, i, 1,2,3) according as i.::;; 3 or i > 3. If one Ilf the dice is selected randomly and rolled twice, then prove that the probahility of obtaining a double-six is -h. Further, given that a double-six has been observed, prove that thl' probability of the selected die being D6 is i. How is this probability altercd if it is known that D3 is a biased die such that, though it is equally likely 10 show a three or a number greater than three, the probability of it showing a six is twice that of obtaining a four or a five? 37 On receiving an external stimulus, a certain kind of biological particle gives rise to a progeny of r (0.::;; r'::;; n) particles with probability (~)p'q" " where p and q are positive parameters such that p + q = 1. The original particle then dies. The particles, if any, constituting the progeny act independently of each other and are subjected to another external stimulus, till' behaviour of these progeny particles being the same as that of the original particle. If Xl and X 2 are random variables denoting the number of particles in the progenies generated after the first and second stimulus respectively, prove that the joint probability-generating function of Xl and X2 is E(OflO~2) == G(Ol> O2) = [q + pOl(q + p02)/I]/I.
Hence, or otherwise, show that (i) P(X2 > 0) = 1- q/l (1 + pq/l-l)/I; (ii) E(Xl + X 2) = np(1 + np);
and
(iii) var(Xl + X 2) = I1pq[np + (1 + np f].
38 A system can be in anyone of sixteen states which may be represented as the equal cells of a square divided by lines parallel to its sides. Initially the system is equally likely to be in anyone of the states. The sixteen states are of three types conveniently denoted as A, Band C, and their distribu-
SUPPLEMENT • J1
11 0
323
is as given in the diagram below. A B B A
B C C B
B C C B
A B B A
After initial placement, the system is moved n times. At each move, the system is transferred to a neighbouring cell in any direction horizontally, "ertically or diagonally, all such moves being equally likely. \ For n;?!: 1, suppose all' {311 and 'YII respectively are the probabilities that after It moves the system is in an A-type, B-type or C-type cell. Prove that Ihese probabilities satisfy the following difference equations: a" =!{311-1 +!'YII-l
{311 = ~all_1 +~{311-1 +!'YII-l 'Y" =1all -l +~{311-1 +i'YlI-l'
Show that the solution of these difference equations is all =!-i{3",
where
10 [ 13 (-9)"] .
{3"=21 1- 90 40
Hence deduce the probabilities that after n moves the system will be in a specific A-type, B-type or C-type cell and verify that, as n ~ 00, these probabilities are in the ratio 3: 5 : 8.
39 Two players A and B contest a series of games. At each trial the probabilities of A and B winning a game are p and q respectively (pi 0, qrO), and there is a non-zero probability r for it to be drawn. It is agreed Ihat the winner of the series will be the player who is the first to win four more games than his opponent. Prove that, for pi q but irrespective of the value of r, (i) the initial probability of A winning the series is 1 1 + (q/p)4 '
and (ij) the probability of B winning the series when he is exactly 4) games behind A is
11
(0 ~ n ~
1-(p/q)4-11 1- (p/q)8
Hence deduce the values of the probabilities (i) and (ii) when a game cannot be drawn, the other conditions for the contest being unaffected.
40 A self-service store has a stock of nt + n undated one pint bottles of milk of which n are fresh and 111 a day old, and it may be assumed that these
324
EXERCISES IN PROBABILITY AND STATISTICS
are placed randomly on the service counter of the store. If a cu selects k (k < Ilt and n) bottles of milk, find the probability that she ~O~lcr fresh milk bottles. as
In
f (v) (In n+~ -:-k k)j(1lt n+ n), n-v-k>O, r
r=O
r
and that she gets no fresh milk is
f (v) (111 + nn~ v- k)j (111 +
r=O
r
r
11).
11
Hence, or otherwise, prove that the conditional probability that the secor I customer obtains k bottles of fresh milk when it is known that at least one :'r her bottles is fresh is I
f
r=O
(v)(I11+~-:-k)j[(m+n)_
nr k
r
n
f
r=O
(v)(m+n~v-k)]. r
nr
41 In a game of chance, a player can score one, two or three points at " trial with constant probabilities q, pq and p2 respectively, where p + q = 1. in repeated trials of the game, Un denotes the probability that the player< total score will be n points at some stage, then show that Un satisfies a third-order linear difference equation with the initial conditions Uo = 1 and Ul = U2 = q. Hence deduce that
Ii
1 [1+(-P)n+l{~Sinn71'/3-(1+P)COSn71'/3}], for 11;;;:0, (1+p+p2) ,,3 42 A finite population consists of the N integers 1,2,3, ... ,N, where N is unknown. A random sample of n integers is taken from this population without replacement, and it is observed that x is the largest integer in the sample. If X is the random variable denoting the largest integer in samples of n obtained from the population, then prove that the probability distrihution of X is given by the set of equations Un
=
P(X=x)= e=~)j(~,
for
n~X~N.
Use this probability distribution to show that for any integer r the rth factorial moment is E[(X + r-1)
Hence verify tliat
(n+r).N. T=(n+1)X 1 n
is an unbiased estimate of N, and that
1)(N -n) n(n+2) .
( ) _ (N +
var T -
SUPPLEMENT
325
[Note: It may be assumed that
f
x="
(X+k)= (N+k+1) l1+k l1+k+1'
ror all integer values of k ~ -1.] 43 A common method of making a saving for the future is by means of periodic payments, usua!ly at equispaced intervals of time. These payments accumulate compound Interest at a fixed rate, and the whole amount is payable after a stipulated number of years. An investor makes an initial payment of £R, and ag.r~es to pay every subsequent year increasing amounts R,2R, 3R, .... In addItIon to these payments, the investor has probabilities pand 1 - p of contributing £2R or £R towards the annuity at the beginning of each year after the first. Prove that the expected amount accumulated after 11 years is
R
SIt =2 [{I +(1 + p)p}{(l + p)" -1}+ p2(1 + p)" - np], p
where the annual percentage rate of interest is lOOp (0 < P < 1). Hence show that, irrespective of the value of p and for 11 large compared with IlP, SIt
Finally, if under the above conditions the investor desires a total return of A, deduce that
1
n >-log[Ap2/R(1 + pf]. p
44 In The Wind on the Moon by Eric Linklater (Macmillan; 1944), the Palfrey sisters, Dinah and Dorinda, have in all the sum of 16 shillings and 11 pence to pay for the unusual services of the barristers Hobson and Jobson. In offering the sum to the barristers, Dinah thinks that it would be rather difficult for them to divide the money equally between them. However, Mr Jobson thinks otherwise and declares "I see there are twice as many half-crowns as shillings and the number of half-pennies is the same as the number of shillings and half-crowns added together, including the very shining bob with the head of Queen Victoria on it. Then if you add to that (that is, the total number of half-crowns and halfpennies) the total number of two-shilling pieces, the total is one and a half times the number of pennies. Now let me see ... " "There are two single shillings", said Mr Hobson, "and if you multiply the number of half-crowns by that you will get the exact number of pennies. It's perfectly easy, Jobson. We divide the 16 shillings and 11 pence into two exactly equal halves, by taking 11 coins each. There are no complaints and nothing left over." Determine the individual face values of the 22 coins offered in payment to the barristers. Hence show that if the 22 coins are randomly divided into two groups of 11 each, then the probability that the resulting division of money will be equal is 200/4199-0·0476. It may be assumed that the symmetrical distribution of coins is the only solution for equal division of the money.
326
EXERCISES IN PROBABILITY AND STATISTICS
Also, show that the same distribution of the face values of the coi . ns IS obtained even if the number of shillings is not known. [Note: For those unfamiliar with the pre-decimal British monetary syste shilling or a bob was worth 12 old pence, a two-shilling piece 24 old P:' a IICr and a half-crown 30 old pence.]
45 (i) Define (x), the distribution function of the unit normal variable X and then show that for any x > 1 . x
1 Jexp[1 -1) +<1> (x-2+ 1) -2 (t2-1)2J dt= (X-2-1.
2J(2'Tr)
-x
(ii) The parliament in Ruritania consists of 2N + 1 members. On . non-party motion, a free vote is proposed. A pressure group consiS\~ of 11 members who agree amongst themselves that they will all VOir in the same way, but it may be assumed that the remaining members will vote independently, and each is equally likely to vote for o~ against the motion. Assuming that there were no abstentions at the actual voting, determine the probability that more than half the total votes will be cast in the same way as those of the pressure group. Jr N is large compared with 11, use a suitable approximation to show that, for fixed N, this probability is an increasing function of II. Further, assuming that the effective range of a unit normal variable is from -3 to 3, prove that it is practically certain that the decision on the motion will go the same way as the voting of the pressure group if 11 ;;;, ~[J(8N + 13) - 3].
46 A population of wild animals consists of N members, where N is a large but unknown number, and a biologist is interested in obtaining an estimate of N. He takes a random sample of W animals, marks and then releases them. After this, he starts capturing the animals randomly and noting whether they are marked or not. He continues this sampling without replacement till this second sample contains exactly w (fixed and preassigned) marked animals. The sampling is discontinued and it is observed thai the total sample size is n. If X is a random variable denoting the size of the sample, prove that P(X=I1)=
(~=:)(:=~)/(~),
for
w~X~N-W+w.
By considering P(X = n) as a function g(N) of the unknown parameter N, show that g(N) g(N-1)
(N -11)(N - W) N(N- W-I1+W)'
Hence deduce that the maximum-likelihood estimate of N is the largest integer contained in IlW/w. Also, assuming that E(X) = w(N + l)/(W + 1), evaluate a simple unbiased estimate of N. [Note: It may be assumed that no change in the animal population occurs in between the taking of the two samples.]
SUPPLEMENT
327 47 A player has a constant probability p (>1) of winning in an independent trial of a .game of c?ance. The sta.k~ for each trial is a shillings. If the player wins a trIal he receives (a + x) shIllIngs (x> 0), but he loses his stake money ~J1 the contrary event. The player decides to play this game with the proviso :hat he would stop play immediately after his first failure. If S is the random variable denoting the net gain of the player at the end of his play, prove that px
E(S)=--a; q
px 2
var(S)=-2 , q
p+q=1.
Also, show that the player can expect to gain at least a shillings at the end of play if x> a/po
48 Small electrical fuses are mass-produced and then packed into large boxes by an automatic filling machine. The number of fuses put into a box by the machine cannot be controlled completely; and in order to ensure that no customer is a loser, the filling machine is so adjusted that it puts at least n [uses in a box. It may be assumed on empirical considerations that the actual number of fuses in a box is a discrete random variable X having the geometric distribution with point-probabilities P(X = x) = (1- p)px-n,
for
X~
n,
where p is a positive parameter such that 0 < P < 1. Given a randomly selected box, the probability that any particular fuse selected at random from the box is usable is a constant 0, where 0 < 0 < 1. If a box contains exactly x fuses, prove that the probability that it contains exactly r usable fuses is P(r I X = x) =
(:)0'(1- oy-r,
for
0:0;;; r:O;;; x,
and write down the expected number of usable fuses in a box containing exactly x fuses. Hence, or otherwise, show that the expected number of usable fuses in a randomly selected box (whatever the number of fuses it contains) is E(r) = nO + pO/(l- p).
49 A manufacturer of a breakfast cereal introduces a coupon scheme as a sales promotion campaign. A set of coupons consists of It (~2) different ones and one coupon is placed in every packet of the cereal. A complete set of coupons can be exchanged for a stock of spring bulbs. A housewife buys one packet of the cereal every week. Assuming that it is equally likely for her to find anyone of the n coupons in a packet, prove that the probability that she will have to wait r (~1) weeks before obtaining the (x + l)th coupon when she already has x different coupons is P(r I x) =
(1-~) (~r-l,
for
1:0;;; x:O;;; n-1.
Hence show that, for any given x, the expectation of r is It/(Il- x), and the expected total number of weeks required for the completion of one set is
328
EXERCISES IN PROBABILITY AND STATISTICS
50 (i) If X is a random variable having the Poisson distribution with /L, and a, (3 and k are non-negative integers such that 0 < k s: l11ean prove that '" Q <:" I~, 13
L r(klp(X = r) = /L kP(a -
k os; X os; (3 - k).
r=a
(ii) A photographic dealer normally sells a standard roll of colou 1'1 for 42 new pence, but in order to encourage sales he deck[ ,1111 allow a discount to customers who purchase more than one rOI~~ III time. To a customer who buys r (2 os; r os; 5) rolls at a time the
where a = P(Oos; X os; 5); b=P(1OS;XOS;4); c=P(Oos;Xos;3). Henn' deduce that var(Z) = 36(1- ~~+e~~b + /Lc)
[6(\~a:_~ /Lbr
What is the average return per roll of film sold under the schellle of the dealer?
51 The attendance records of workers at a large factory kept for a IOIl~ time show that the probability of a worker being absent for r (;;;'0) days in a year was approximately given by the Poisson distribution with mean IJ.. The absence of a worker causes dislocation in the production schedule which results in a loss to the factory; and it was estimated that, on the average, if a worker was absent for r days in a year the net loss to the factory was £a/'(r+ 1) e-13r, where a and (3 are positive parameters and (3 is small. Prove that the expected loss (in pounds) to the factory due to absenteeism of a worker in a year is a/L(/L e- 13 + 2)exp[ -(/L + (3) + /L e- 13 ], Hence show that if /L and (3 are such that 1 < /L < el3, then the expected loss must be less than 3a.
52 The Poisson distribution is occasionally referred to as a distribution describing the occurrence of rare events. Explain this remark. A commercial secretarial organization charges Sa pence for cutting a foolscap stencil. The price includes the cost of the stencil and the remuneration of the typist, and the residual profit of the organization is a fifth of the charge made, provided the typist makes no typographical errors. However. if an error is made, then the profit of the organization is diminished by the additional cost of correction, On empirical considerations, it is known thaI the cost in pence of correcting r errors on a stencil is 2r(3r + 2)/(r+ 1), for r;;;' 1, It may be assumed that, on the average, the typist makes /L errors per
SUPPLEMENT
329
.Ieocil where /L is.lsmaldl. Shuggesht a sUhitable model for the distribution of ors on a stenci an t en s ow t at the expected cost in pence of :rtrrecting the errors made on a stencil is .0 2[3/L -1 + /L -1(1-e- lL )]. Ileoce show that if the organization wishes to obtain, on the average, a oftt of A.a pence per stencil, where A. is a constant such that 0 < A. < 1, and \ assumed to be sufficiently small for terms involving /L 2 to be negligible, ;hell. as a first approximation, the organization will attain its objective if 5/L
a=I_A. .
<3 A stationery store plans an advance order of a standard line of next :ear's calendars which are usually sold in the pre-Christmas season. The order can be placed for any number of calendars, and the cost per calendar 10 the store is two shillings and sixpence. The store expects to sell the 'alelldars in the pre-Christmas season at a (1 < a < 2) times the cost price, ~ut the residual stock will have to be disposed of in the January sale at a fraction (3 (0 < (3 < 1) of the cost price. The past experience of the store ,hoWS that the pre-Christmas sales of this line of calendars are Nx, where N ~ a positive integer and x is a random variable having a Poisson distribution lI'ith mean /L and distribution function F(x). If the store orders Ny calendars, show that the expected profit in pence of the store is y G(Y)=30N[(a-(3)
,,~u (x
y)p(x)+(a-l)YJ,
II'here P(x)=e-ILIL"/x! Hence, by considering the behaviour of the finite difference G(y + 1)G(y), prove that the optimum size of the stock which the store must order to maximize its expected profit is Nyu, where Yu is the smallest integral value of y for which a-I a-(3
F(y)~--.
54 A government wishes to control immigration by restricting the entry of foreigners to an annual intake of a persons. (It may be assumed that the immigrants enter the country at the commencement of each calendar year.) Suppose initially, at the start of the first calendar year, the total population of the country is a constant N. However, due to the natural changes in the population, it is presumed that the actual number of persons in the country at the end of the year will be Nx, where x is a random variable such that its probability density function is A. e-A(x-l),
for x ~ 1 and A. > 1 is a parameter.
Accordingly, at the beginning of the second calendar year the total population, including the annual quota of immigrants, is another random variable }'2=a+Nx.
Prove that E(Y2)=a+N(I+A.)/A.. In general, suppose that the expected population, including the immigrant quota, at the beginning of the rth (r ~ 3) year is 1 + A.) E(Yr-l)' E(Yr) = a + ( -A.-
330
EXERCISES IN PROBABILITY AND STATISTICS
Use this equation to verify that the expected population at the beginni the (k + l)th year is ng of a[
(1 + "f - AkJ
Ak -
1
+N
(1 + A)k A .
Hence deduce that, if the annual immigrant intake a = N/ A, the popular will attain IJ. times its initial size at the beginning of the (k + l)th year whe101J . , re . fi es the equation k satIs 2(1 + A)k = (IJ. + l)A k. If A is large, show that an approximate solution of this equation is
1)
IJ. + . k =(l+A)log ( -2-
55 The annual profit of a steel company depends upon the demand x and the two scaled parameters a and (3 representing the capacity of the plant and the selling price of steel respectively. It is known from past experience that, for given a (>1) and (3 «1), the profit increases with rising demand However, the effect of an increase in price is to decrease demand due to th~ competition of imported steel; consequently, for given a, the total realizable profit tends to decrease somewhat because of the additional cost of holding unsold stock. Accordingly, it is suggested on empirical evidence that a reasonable approximation of this dampening effect of price increase on profit is a factor exp[-(3(a - x)], which proportionately reduces a parabolic increase in profit with rising demand. It may therefore be assumed that, on a suitable scale, the profit of the company is T= aX(l + (3x)
e-!Ha-x).
Hence, assuming that the demand is appropriately scaled so that x is a normally distributed random variable with zero mean and variance (3, prove that the expected profit is E(T) = a(32«(33 + 2)exp[-(3(a - (32/2)]. Finally, suppose (3 is fixed by a retail price agreement amongst the steel companies, and a company has the freedom to determine the production capacity of its plant for maximizing the expected profit. Show that under such a restrictive price arrangement the company can obtain maximum profit when a = 1/(3.
56 A factory produces corned beef for export which is despatched in large consignments of 12 oz. containers to customers overseas. However, due to a recent failure in the sterilization of the product, the government has imposed the restriction that a consignment should only be allowed to pass for export if a random sample of 100 containers from it is found to be completely free from infection. It is suspected that, despite improvement in the sterilization process at the factory, there is an a (small) per cent rate of infection in the consignments offered for inspection. If it may be assumed that the size of the consignments is large compared with the sample size so that the effect of sampling from a finite population is negligible, prove that the probability of a random consignment being passed as free of infection is approximately
SUPPLEMENT
331 Hence deduce that the probability that of a batch of ten consignments least three will be rejected on inspection is
_fl.
:1
1-
L 2
w=O
(10) (l-e-"')W(e-"')lO-
w•
W
If, over a period of time, this probability is assessed to be 1· 5 per cent, verifY that the rate of infection still present in the consignments is approximately five per cent. [Note: e- o.os = 0.95 correct to two decimal places.]
57 A printing house proposes to pay £a as a standard charge for proofreading a book irrespective of the number of errors detected by the proof-reader. If no errors are detected in the proofs, then the proof-reader is paid another £ e-Aa, where A> 0 is a parameter. On the other hand, if the proof-reader detects errors in the proofs, then he is paid £pr-1a if he detects r errors for 1::;;; r
P(x = 0) = l-e- A ; P(x=r)=(I-e- A )e- r>-,
for l::;;;r
Assuming that the proof-reader detects the errors present in the proofs, shoW that the expected amount payable to him is e-A(I-e-A)(2-p e- A)] a + I-pe - A '
[1
and hence deduce that this amount can never exceed 7a/4.
58 A retailer obtains his supplies of a small machine component from the manufacturer in large batches, each batch consisting of M components. Although the production is under statistical control, there is a proportion e of defective components in each batch such that e may be regarded approximately as a continuous random variable with the probability density function 60(1- 0) in the range 0::;;; e::;;; 1. The retailer sells each component for one shilling but, to attract custom, he agrees to refund a shillings (1 < a < 2) to a customer for a defective component supplied. To protect himself under such a guarantee system, the retailer takes a sample of N (fixed and «M) components from each batch supplied to him by the manufacturer for destructive testing, and refuses to accept any batch containing n or more defective components. For any given 0=0, find the conditional probability of the retailer accepting the batch after inspection, and then show that his expected net income in shillings from such a batch is g(n, 0) = (M - N)(I-aO)
~t~ (~W;(I- O)N-x.
Hence deduce that the expected net income of the retailer per batch for variation in e is g(n) =
(M - N)(n + 1)(2) 2(N+4)(4) [6(N+2)(N +4-a)-2(2n + 1){(N + l)a + N +4}+
+ 3an(n + 1)].
332
EXERCISES IN PROBABILITY AND STATISTICS
Finally, assuming that N is sufficiently large for terms involving N-I be negligible but 11/N is finite, show that g(n) is a maximum for 11::: N;o approximately. Q
59 A factory produces packets of margarine and the nominal weight of th contents of each packet is one lb. However, due to random fluctuations .c the automatic weighing machine, the actual weight of the contents of I~l packet is a normally distributCi:d random variable x with mean IL a ~ variance u 2 (IL »u). It is also known from the past experience of the facton( y that . the cost in shillings of producing a packet containing x lb of margari,r~ IS
0: + (3x - (x 2 -
( 2)
e --v(x-/-L)/u,
where 0:, (3 and 'Yare small positive parameters. Prove that the expectcd cost of a packet of margarine is a
+ (3/1- - (/1- - 'YU f e _~-V2.
An economic crisis forces the government to devalue its currency and consequently, there is an increase in the import price of edible oils used ir; the production of the margarine. To offset a part of the increase thc manufacturer makes certain small economies in his fixed costs and also establishes checks to lessen waste. As a result of these measures, thc parameters of the cost function are expected to become 0:' «0:), (3' (>~) and 'Y. The manufacturer aims that, on the average, there will be no increasc in his cost of production; and he plans to achieve this by suitable alterations in the parameters of the weight distribution. Assuming that the new weight distribution is normal with mean /1-' and variance U,2, and /1-' is so chosen that 0: + (3/1- = a' + (3' /1-', prove that there will be no change in the expected cost of production if /1- -/1-' = 'Y(u-u'). Hence show that the production of packets produced by the new system which are less than the nominal weight is ['Y{-(O: - 0:') + «(3' - (3)/1- + (3'(1-/1-)J (0: - a') - «(3' - (3)/1- + 'Y(3' u ' where (z) is the distribution function of a unit normal variable z. 60 It is known from the past experience of a publishing house that a small proportion of typographical errors remains undetected by the usual methods of proof-reading. In order to improve upon the existing procedure, the publishing house agrees to an additional proof-reading of a book with the proposal that the proof-reader will be paid (i) £0·25 for a page on which he detects no error; (ii) £1·00 for a page on which he detects a single error; and (iii) £1·25 for a page on which he detects more than one error. The probability that a word on a page is typographically incorrect is a small number a and, on an average, there are 11 words on a page. If /I is sufficiently large compared with 0: but 110: = /1-, a finite number, give without proof the probability that there are exactly r (~O) errors on a page. Hence,
SUPPLEMENT
333
assuming that the proof-reader detects the errors present, show that the eXpected payment per page to the proof-reader is £~[5 -
(4+ IL) e- IL ].
The proof-reader is unwilling to accept the suggested rates of payment because he realizes that it is unlikely that many pages will contain one or ptO re errors per page. As a counter-proposal, the proof-reader asks for an arrangement which would give him an expected payment of £1·00 per page by raising the amount £0·25 in (i) when the amounts £1·00 and £1·25 in (ii) and (iii) remain fixed. Prove that the increased payment that the publishing house should make for each page on which the proof-reader finds no errors, in order to meet his demand is £M5 + IL -e- IL ].
61 A manufacturer of a duplicating machine rents out the machine to cuslomers on an annual basis. The annual rental for a machine is fa, a fixed charge, plus an amount which depends upon the number of copies made in a year. Accordingly, if x copies are made on a machine in a year, then the lolal rental payable (in pounds sterling) is T(x) === a
+ bx(1-tx),
where b is another constant. If it is known from empirical considerations that x can be regarded as a continuous random variable with the probability density function (A + 3)(A + 2)x (1+X)A+4 ,forx;;a.O,
where A > 1 is a parameter, prove that for r;;a. 0 E(x r ) = f(r+2)f(A - r+2)/f(A +2).
Hence deduce that E(T) =
a
+ (2A -1)b
A(A+1)'
The manufacturer wishes to so arrange the rental of the machines that, on the average, he will obtain a rental income of at least £Aa per machine. Show that the manufacturer will attain his objective if
a
2A-1
b
A(A -1)'
-<---::-2
62 In a certain welfare state, a person pays an annual contribution for social security based on his income x which is greater than a, the minimum individual income in the population; the amount of the contribution is A(x - a), where A is a parameter such that 0 < A< 1. The probability distribution of the incomes of individuals liable to pay the social security contribution is log-normal with the probability density function _1_(x-a)-lexp-_1_[log(x-a)-lLf, for x>a, u.,fi;,. 2u 2 Where u and IL are parameters in standard notation. Prove that, on an average, the contribution paid by an individual is A exp(1L + !u2 ).
334
EXERCISES IN PROBABILITY AND STATISTICS
As a concession to income earners, the government proposes that 0 I persons with incomes greater than (3 (>a) should be required to pay ~ y social security contribution at the rate of A(X - (3). Show that the aVer he contribution now payable is A times age [1-<1>(8 - 0')] e fL +¥>"2 - ({3 - a )[1- <1>(8)], where 8 == [log({3 - a) -IL ]/0' and <1>0 is the distribution function of th standardized normal variable. Hence deduce that the relative change in the average social security contribution is e <1>(8 - 0') + [1- <1>(8)] exp 0'(8 -!O').
63 An owner of a large Edwardian mansion has the choice between tw methods of heating the residence. He can either continue to use solid fuel i:: the existing fire-places or replace them by thermostatically controlled gas fires. It is estimated that the annual consumption of solid fuel would be N tons at a cost of £a per ton. On the other hand, the initial cost of installing t~e gas fires is substantial and it m~y be discounted as an annual cost of £(l Smce the gas fires are thermostatIcally controlled, the consumption of gas would depend upon the daily temperature distribution as recorded within the house. For anyone day, the gas consumption in therms may be regarded as a random variable X with a probability density function proportional to e- Ax2 . X3 for X>O, and A > 0 is a parameter. Prove that the average consumption of gas per day is N w/ A therms. Hence, assuming that the price of gas is £ 'Y per therm and that, on the average, there are 300 days in a year when heating is at all required, determine the expected annual cost of heating the house by gas. Also, show that, on the average, solid fuel heating will be more economical as long as 'Y
aN-{3
J>.
22SJ";'
->---
64 A refinery, located at a seaport, has a capacity of processing N gallons of crude oil during a week. The refinery is supplied with crude oil by tankers which arrive at random intervals of time; and if the refinery receives no fresh crude oil during a week then it is scheduled to process only aN gallons, where 0 < a < 1. The amount of crude oil received by the refinery during a week is a random variable X, and as X increases the scheduled processing capacity of the refinery also increases. However, since the capacity of the refinery is limited, it may be assumed on empirical considerations that this increase dies off exponentially. It may therefore be assumed that if the amount of crude oil received by the refinery in a week is X, then the scheduled processing amount is T = N[l-(l-a) e- 13x ], where (3 «1) is a small positive parameter. If the probability density function of the probability distribution of X is f(x)
=
x
e-~x2
for X;;:. 0,
prove that E(T) = N - N(l-a)[l-J2;. e!13 2{3{1-({3)}],
335
SUPPLEMENT
where. ct>(.) is th~ distri?ution function ~f a unit normal variable. Hence, aSSumIng that (3 IS sufficiently small ~or (3 to be negligible, show that in any one week the amount of crude oIl scheduled for processing cannot be greater than the expected amount if
0 1 log (7T X";-I3· 1- s e12) 2"
•
6S A continuous random variable X has the probability density function proportional to x"'-I/(l+(3x"')2",
for X;;oO,
where a and (3 are positive parameters and n > 1. Prove that if (2/1-1)a - r > 0, then the rth moment of X about the origin is given by
E(X') = 2n-l (3'/""
B(~+
12n -1-~) a
a'
in standard notation for the beta function. Also, show that the stationary values of the probability density function of X are the roots of the equation a '" ,..x
=
a-I (2n-l)a+l .
Hence deduce that for a> 1 the probability distribution of X has a mode at the point [
a-I
]
1/",
X= {(2n-l)a+l}(3
.
66 Explain, very briefly, the concept of conditional probability. A continuous random variable X has the beta distribution of the first kind with the probability density function 1 - - - x l1l - l (l-x) B(m,~
,
for 0..; X..; 1
,
where 111 is a positive parameter. Given a a positive number less than unity, prove that P(X ..;a) = a "'(111 + l-l11a).
Given another positive number (3 < a, prove that . ((3)111 111 + 1 - m(3 P(X"; (3 I X..; a) = - . 1 . a 111+ -l11a
67 A continuous random variable X has the probability density function ![1 + h(l- x 2 )], for -l..;X..; 1, where A is a parameter having some value in the interval (-1,1). Prove that for any non-negative integer r
E(X')=.!. [_1_{1+(-lY+2}+ 2A {I + (-lY+3}]. 2 r+ 1 (r+2)(r+4)
336
EXERCISES IN PROBABILITY AND STATISTICS
Hence verify that 2A
var(X) =
E(X) = 15 ;
75-4A 2 225
Also, prove that the distribution function of X is
( 21[A4
A
1 2)] F x)=- l--+x+-x 2( 1-2:x
2
.
Hence show that the median of the probability distribution of X is a rc' I root of the equation a
A(x 2 -1f-4x = 0, and that this root is > or <0 according as A> or <0.
68 A continuous random variable X has the probability density function proportional to x 2 /(1 + x 4 ), defined in the range -00 < X < 00. If k denotes the proportionality factor, prove that k (2r+3 1-2r) E(X2 r)=2 B - 4 - ' - 4 - '
. provided 1-2r>0,
in standard notation of the beta function. Hence deduce that k = .f2/7r and that none of the even moments of X exists. Also, show that the turning values of the probability density function of X are the roots of the equation
X4=1. Hence prove that the probability distribution of X is bimodal and determinc its modes. Finally, show that the points of inflexion of the probability distribution of X are the real roots of the equation
3x 8 -12x 4 + 1 = 0, and determine these roots explicitly. [Note: It may be assumed that f(t)f(l- t) = 1T/sin 1Tt.]
69 A continuous random variable X has the probability density function ex(3x",-l f(x) = (1 + (3x"'f' for X;;;'O, where ex and (3 are positive parameters. Prove that if ex > r, then the rth moment of X about the origin is given by E(xr) =_l_B(l +..!:. (3r/'" ex '
1-..!:.)
a'
in standard notation for the beta function. Also, for a> 1, show that the mode of the probability distribution of X is at the point [ ex -1 ]11'" x= (3(a+1) .
SUPPLEMENT
337
Finally, prove that the median of the probability distribution of X is at the point
flen ce , for ex> 1, infer the sign of the skewness of the probability distribution of X. 10 (i) For any two positive numbers ex and v, the incomplete gamma function is a ra(v)=r;v)
J
e-xx,,-l dx.
o
If
v> 1, then show by suitable integration by parts that e-aex,,-t ra(v-1)- ra(v) =
r(v)
(ii) For any r;;;:: 0 and 0> 0, prove that II
[(r, 0)=
Je-
4x2 x r dX=2(r-1)/2rA(r~ 1)r(r~
1),
o
where A =!02. (iii) A continuous random variable Z has a non-normal probability distribution with the probability density function 1 [Z4_6z2+3] fu e- tz2 1+ n ,for -oo
P(Z:SO; 0) = <1>(0) + e-!1I2 8 (3- ( 2 )
,
nfu
where <1>(.) denotes the distribution function of a unit normal variable and 8 is a positive constant.
71 A continuous random variable X has the probability density function porportional to (ex + Ix!) exp- ~ lxi, defined for -oo<X <00, where ex and ~ are positive parameters. Evaluate the proportionality factor and then verify that for any integer r > 0 E(X2r )
= (ex~ + 2r+ 1)(2r)! (ex~ + 1)~2r
Hence deduce that 'Y2, the coefficient of kurtosis of X, is 24
'Y2 = 3 -
(ex~ + 3)2 '
so that, irrespective of the values of ex and ~, l< 'Y2 < 3. 72 A continuous random variable X has the probability density function proportional to (ex+x4)e- X2 defined for -oo<X
338
EXERCISES IN PROBABILITY AND STATISTICS
posItIve parameter. Determine the proportionality factor and verify thai E(X) = O. Also, show that the (2r)th central moment of X is 1L2r
(X) = [4a + (2r + 3)(2r + 1)] r(r + !) (4a + 3)-/;
Hence deduce that X has unit variance if a =~. Prove that in this cas mode of the distribution of X is at the origin and the maximum ordina~ th~ the distribution of X is 3/(4";;). Also, show that the coefficient of kurtC (!f of the distribution of X is OSI, I'z(X) = -i. Compare these values with the corresponding ones for a unit normal distribution and comment on the differences observed.
73 The distribution function of a continuous random variable X having th . "logistic" distribution is defined by the equation ~ P(X:os; x) = (1 + e- x )-l,
for -00 < X < 00.
Evaluate the probability density function of X. By using the transformation y = (1 +e-X)-l, or otherwise, prove that the moment-generating function of X about the origin is Itl< 1. Hence deduce that 1'1 and 1'2, the coefficients of skewness and kurtosis of X are 0 and 6/5 respectively. . Also, show that (i) the semi-interquartile range of X is loge 3; and (ii) the points of inflexion of the probability density function of X arc
74 A continuous random variable X has the probability density function proportional to x 4 /(l + x 2 )(1 + X4), for -00< X < 00. Prove that the proportionality factor is 2/7T. Show that the turning values of the distribution of X are the roots of the equation X6
-x 2 -2=O.
Hence deduce that the distribution of X is bimodal and, assuming that the modes lie in the neighbourhood of the points x = prove that a more accurate approximation for the modal values is x = ± 35/23.
±J3i}'
75 A fluid flows continuously at a constant rate of Q molecules per minute into and out of a reactor. The reactor is said to be "well stirred" if the molecules inside it are always mixed in such a manner that the probability that any given molecule will leave the reactor during an infinitesimal time-interval Sx is approximately Sx/(3, where (3 == V/Q, V being the volullle of molecules inside the reactor. It is then known that the time for which a molecule remains in the reactor is a random variable X having a negative exponential distribution with mean (3. However, if two well-stirred reactors, each containing V molecules, are connected in parallel, then the entering fluid splits between them. If the total flow rate is 2Q molecules per-minute and a fraction a (O
SUPPLEMENT
339
loes through one of the~~ rea~to~s a?d the remaining 1 - a through the ~ther, th~n t~e probablhty. ~hstnbut.lOn of X is approximately hyper~%ponentIaI, with the probablhty density function f(x)
= 2;2 [e- 2aX/l3+
C:o'Ye-20 - a
)x/13
J,
O~X <00.
prove .that in this case the moment-generating function of X about the origin IS 2[2O'(1-O')-{1-2O'(1-O')}{3t]
Mo () t = ---'--------'-------'--'-~ (20' - (3t)[2(1- a) - (3t] . Hence deduce the mean and variance of X. Also, show that, irrespective of the value of a in its permissible range of variation, E(X) = (3
var(X) ~ {32.
and
16 Evaluate the moment-generating function about the mean of a unit normal variable X. Hence verify that the (2r)th central moment of X is (2r)! I-'- (X) - - -
2r
- 2r. r! '
for r;30.
Suppose Z is a continuous random variable with the probability density function proportional to e-!z2[1+O'z(z+1)],
for -oo
where a is a positive constant. Prove that the odd and even moments of Z about the origin are ,
(Z)
a
(2r+2)!
1-'-2r+1
=1+O'·2 r+ I .(r+1)!'
1-'-2r(Z) =
1+(2r+1)O' (2r)! 1 . -2r" +0' . r.
and for r;3 O.
Hence determine the mean and variance of Z. Also, show that both E(Z) and var(Z) are increasing functions of a but that, as a ~ 00, their limiting values are 1 and 2 respectively. 77 For a continuous random variable X having the negative exponential distribution with probability density function A e-AX,
for X;30,
obtain the cumulant-generating function of X. Here deduce that the rth cumulant of X is (r-1)!
Kr(X) = -A-r-
, for r;31.
Suppose e- Ay is regarded as proportional to the point-probabilities of a discrete random variable Y which takes all the non-negative integral values 0,1,2, .... Show that the probability-generating function of Y is 0(8) =E(8 Y ) = (1-e- A)/(1- 8 e- A).
340
EXERCISES IN PROBABILITY AND STATISTICS
Hence determine the cumulant-generating function of Y and verify thaI KI(Y)=1/(e A -1);
and
K2(Y)=e A/(e A -1)2.
Finally, if >.. is sufficiently small for>.. 2 to be negligible, prove thaI
KI(X)-Kl(Y)=~ (1-~)
and
K2(X)-K2(Y)=A
78 A standard pack of playing cards has four aces and 48 other card, l pack is well shuffled and then cards are dealt out one by one face up~. hl, until the second ace appears. If X is a discrete random variable denotin number of cards dealt out before the appearance of the second ace pg. Ill' that ' 111\\.
clrt
P(X = r) =
(~)C~81)(523_ r) /
(5r2) , for 1 ~X ~49.
Also show that P(X = r) satisfies the recurrence relation P(X=r+l)=
(r+ 1)(49- r) ( ) .P(X=r), for 1~r~48. r 51- r
Hence verify that P(X = r) is an increasing function of r only if r"", 16. Explain, very briefly, how the above recurrence relation can be lIsed obtain the numerical value of P(X ~ 5).
10
79 Explain, very briefly, the concept of unbiasedness of an estimate in Ihl' theory of estimation. A continuous random variable X has the probability density function x A f( x) = _1_ 6>..4 e- / .X3, for X
~ 0,
where>.. is a positive parameter. Prove that E(X) = 4>..
and
var(X) = 4>.. 2.
If Xl> X2, •.. ,X" are 11 independent observations of X, show thaI maximum-likelihood estimate of >.. is
llll'
A=~i,
where i is the average of the Xi' Hence deduce that an unbiased estimate of >.. 2 is ni 2 /4(411
+ 1),
whereas the estimate i 2 /16 has a bias of O(n- I ). 80 A continuous random variable X has a normal distribution such lhal both the mean and variance of X are equal to 0, unknown. If XI' Xb ','" x" are independent observations of X, show that the equation of 8, lhl' maximum-likelihood estimate of 0, is e(e+1)=i 2+s 2,
where n
ni=Lxj j~l
n
and
ns 2 =L(xj-if. i~l
SUPPLEMENT
341
~eoce deduce that
8 = ~({1 + 4(.XZ + S2)}! -1] ]lIII that, for large samples,
20 2 var(O) = n(28+ 1)' A
11 Explain,. ver'! briefly, the concept of unbiasedness in the theory of tatistical estimation . . A continuous random variable X has the probability density function
1
f( x) = -3 x 2 e- x/ fJ for X;a.O. , 20
prove that for any r;a. 0
E(xr) = !orf(r+ 3). Hence verify that E(X)=30;
Given that XI> X2, ••• , X" are independent realizations of X, prove that the maximum-likelihood estimate of 8 is 8=li,
Ilhere i is the average of the n observations. Also, determine var(O). Two functions 1 " L1 = x~ and ~=li2 4n i=1
L
Jre suggested as possible estimates of var(X). Show that Ll is an unbiased of var(X) but
~Iimate
\0
that L2 is asymptotically unbiased.
82 An infinite population consists of white and black balls, and it is known that p, the proportion of white balls in the population, is rather small. In order to obtain a good estimate of p, an experimenter decides to sample the balls one at a time till he obtains exactly r (fixed in advance) white balls in the sample. (i) If X is a random variable denoting the size of the sample which includes exactly r white balls, prove that P(X=n)= (;=Dprq,,-r,
for X;a.r;
q==l-p.
(ii) Prove that the probability-generating function of X is E(OX) = (pO)'(1-qOrr.
Hence, or otherwise, deduce that E(X)= rIp.
342
EXERCISES IN PROBABILITY AND STATISTICS
(iii) Given r and the sample realization n of X, prove that the ma . )(llllll01_ likelihood estimate of p is r
A
p=-, 11
and that for large samples
p2q
var(p)=-. r
(iv) Also, verify that
E(:=~)=P, so that
p is
not an unbiased estimate of p.
83 In the distribution of the number of police prosecutions per motorist' · over a given . . d 0 f time, ' . 'III a Iarge cIty peno t he f requency 0 f motorists havin prosec~tions is 11" for r ~ 1 ~nd .2::::"=1 nr == N, b~t the number of motori~t: ~ho dId not ?ave a pro~ecutIo~ IS unknow~ owmg to the motorist popul
p,*=Lr.I1.1N r=2
is an unbiased estimate of p, and that var(p,*) =E: [1 +-p,-]. N eIJ.-1 Also, verify that an unbiased estimate of var(p, *) is
2n2+Np,* T2 =
N2
Give any reasons which in your opinion make the estimate p, * preferable to the maximum-likelihood estimate ,1. [Note: It may be assumed that for large samples
p,(1-e-IJ.)2 ] var(p,) = N{l-(p, + 1) e-IJ.}· A
84 Explain briefly the concept of an estimate of a population parameter, and distinguish clearly between unbiased and consistent estimates. Illustrate your answer by suitable examples. The probability density function of the Cauchy distribution is 1 -. 7T
1 , for -co> X
where p, is a parameter of the distribution. If Xl' X2> ••• ,x.. are independent observations of the random variable X, determine the equation of ,1, the maximum-likelihood estimate of p,. Also, prove that the large-sample variance of ,1 is 2/n.
SUPPLEMENT
343
I~ A continuous random. variable X has. ~ normal ~istributi~n with mean (J
,od variance k(J, where k IS a known posItIve quantIty but (J IS an unknown Jd aOleter. If Xl> ~2' •.. ,XII are n independent observations of X, show that ;'Jf equation for (J, the maximum-likelihood estimate of (J, is ·he e(e+k)=.e+s 2,
"
ni=Lxj
and
"
ns 2 =L(xj-if.
i=1
i=1
~ence deduce that
e =![{k 2+4(i 2+s 2)}L k], ad that, for large samples, J
2k(J2 var«(J) = n(2(J + k) A
What is the limiting value of var(O) if, for fixed n, k ~ 0, (J ~ 00, such ,hat k6 ~ /-L, a finite number?
86 Explain, very briefly, the concept of the best estimate of a parameter in ,he theory of linear estimation. Suppose that Xl' X2, ... ,XII are n independent observations such that (J2 E(x,,) = /-L; var(xv ) = , for 1 ~ v ~ n, n-v+1 Il'here /-L and (J2 are independent parameters. Determine /-L *, the leastsquares estimate of /-L, and show that 2(J2 var(/-L *) = - - n(n + 1) Another linear function suggested as an estimate of /-L is
6 T= (
n n+
1)(2
11
n+
1) L (n-v+1) 2xv. v=l
Prove that T is an unbiased estimate of /-L but that, as compared with #.t *,
2(2n+ 1)2 Eff(T) = 9n(n + 1)
8
9
for large n.
87 Explain clearly the meaning of linear regression and state the assumptions underlying the standard tests of significance in linear regression analysis. Suppose that Yl> Y2' ... ,YI1 are independent observations such that E(yJ = a
+ (3(x" - i) + "(x,, - if;
where x" are values of a non-random variable, i = I~=1 xv/n, and a, (3, " (1-0) and (J2 are independent parameters. Prove that, in general, the linear functions 11
a*=
L
yJn
v=l
are biased estimates of a and {3 and find their expectations explicitly. What happens if the x" are equispaced?
344
EXERCISES IN PROBABILITY AND STATISTICS
Also, for general Xv, determine a linear function of a* and (3* expectation is independent of 'Y and find its variance. wh()~t
88 Suppose that Yl> Y2, ... , Y.. are n independent observations such thaI E(y.,) = a + (3(xv - i),
var(y.,) = 0'2,
for v = 1, 2, ... , n,
where the Xv are values of a non-random explanatory variable with av . i, and a, {3 and 0'2 are independent parameters. If a* and {3* a er,lgt standard least-squares estimates of a and (3 respectively, then t;e tht residual is defined as e /Jlh rv = yv-a*-{3*(xv-i),
for v = 1, 2, ... , n.
Prove that the rv satisfy the two linear constraints n
L
,.
rv = 0
L
and
1'=1
(xv - i)rv = O.
1'=1
Also, prove that for any v
n-1 var(r.,) = 0'2 [ n
E(r.,) = 0; where
(x., -
X
i)2]
.
.
L
X=
(x.,-if;
v=l
and
i
var(r.,) = (n -
2)0'2.
v=l
89 Suppose (Xi> Yi), for i = 1, 2, ... , n are independent paired observations such that each Yi is normally distributed with E(Yi) = a
+ (3(Xi -
i)
and
var(Yi) =
0'2,
where a, {3 and 0'2 are the parameters in standard linear regression analysis and i is the average of the Xi> the values of the explanatory variable. If n'; and {3* are the usual least-squares estimates of a and {3 respectively, find (J *, the best estimate of the parametric function (J = a - {3i. Prove that corr«(J*, (3*) =
i.Jn/
(t1 xfY.
Also, determine the value of this correlation coefficient when, in particular, the Xi (i = 1, 2, ... , n) are the first n natural numbers, and then show that, as n ~ co, the limiting value of the correlation coefficient is -!J3. 90 Explain briefly the concepts of unbiasedness and efficiency in the theory of estimation. It is known that, in biological investigations carried out over prolonged periods of time, the experimenters often obtain quite accurate information concerning the coefficient of variation (standard deviation/mean) of the quantitative characteristic under study. This information can be used to plan further experimentation. Since large random variability is a typical feature of biological material, a knowledge of the coefficient of variation can be
SUPPLEMENT
345
Jsed 10 ~btain an asymptotically unbiased es~imate of a parameter, which .as a van~nce smaller than that of ~h~ best lInear unbiased estimate. , In particular: suppose a ~haracten~tIc Y. (su~h as the length of the ears of ~aiZe of a particular type) IS under Investigation, where
E( Y) = IL
and var( Y) =
0'2,
and 0'2 being independent parameters. It may be assumed that v == 0'/ IL is rnown accurately from previous similar experimentation. If Yl> Y2, ... , YI1 JlC independent observations of Y, an estimate of IL is defined by the linear [Odelion i=l
~here
C is so chosen that E(T-IL)2 is a minImum. Prove that C =
1/(11 +v 2 ). Hence show that T has a smaller variance than the sample
Jverage
y, but that E(T) = IL
(1+:2)-1
so that T is an asymptotically unbiased estimate of IL. Also, prove that the tlliciency of T as compared with y for the estimation of IL is (1+
:2y.
••• ,x,. are n independent observations with mean and variance 0'2. Another independent sample of n correlated observalions Yl> Y2, ... ,YI1 is given such that for all i = 1,2, ... , n
91 Suppose that Xl> X2, ~
E(Yi)=IL;
var(Yi) = 0'2;
and
corr(YhYj)=P,
(i=/=i).
Prove that x and y, the means of the x and Y observations respectively, are unbiased estimates of IL and determine their variances. If T== ax + (3y is a linear function of the sample means, find a and (3 such that E(T) = IL and var(T) is a minimum. Verify that this minimum var(T) is 0'2 - [1 +(n -1)p]/[2+(n -1)p].
n
Also, show that (n _1)2p2 ] var[(x + y)/2] = [ 1 + 4{1 + (n -1)p} . min var(T).
Hence, or otherwise, deduce that for n> 1 and P=/= 0 var[(x + y)/2] > min var(T) and comment on this result. 91 If Xl> X2, ••• ,Xm are independent observations of a random variable X having a normal distribution with mean IL and variance 0'2, find the least-squares estimates of IL and 0'2. Hence indicate, without proof, how an exact test of significance based on Student's distribution may be used to test any null hypothesis about IL.
346
EXERCISES IN PROBABILITY AND STATISTICS
Given the 11 sample observations, an experimenter wishes to obt . interval estimate for an additional independent but unknown observat~tn an X. Show how the argument leading to Student's distribution may be IOn or ified to derive the required 100(1-1/) per cent (0 < 1/ < 1) confi:odinterval for the (11 + l)th observation as encc i -t(1/;
1l-1)S~ (11: 1)~x ~i + t(1/; 1l-1)S~ (11: 1),
where x is the unknown observation, i and S2 are the sample mean v~ri~nce. of th.e Xi and t( 1/; 11 - 1) is the 1001/ per cent point of Stude~~~ dlstnbutlOn with d.f. .
11-1
93 Explain clearly, without proof, how a confidence interval may b. obtained for the regression coefficient in linear regression analysis, and Sht c. the underlying assumptions of the analysis. (C In an investigation the distinc~ values of t.he explanatory variable x arc Xl> X2, ••. ,Xn and the correspondmg observations of the dependent variabl' yare Yl, Y2, ... ,Yn, it being assumed that the standard linear regressi{)~ holds between X and y, so that E(Yi) = a + (3 (Xi - i) var(yJ = (J'2, for i = 1,2, ... , 11, and i is the average of the Xi' The experimenter wishes to make k (preassigned and ~ 1) further independent observations of the dependent variable, each observation corresponding to the same given value Xo 1= Xi (i = 1,2, ... , of x. If Yk is the average of these k observations (yet to be made), prove that the 95 per ccnt confidence interval for Yk obtained on the basis of the known (Xi. Vi) i = 1, 2, ... ,11 is
11)
[1 1
A
a+(3(xo-i)-tO.05 xs 'k+-;;+
(X o -i)2J!
X
~
[1 1 (xo-xf]\ X •
~Yk~a+(3(xo-i)+tO.05xS 'k+-;;+ A
where a, ~ and S2 are the usual least-squares estimates of a, (3 and (12 respectively, X=L:t~l (Xi -i)2, and to'05 is the 5 per cent point of Student's distribution with 2 d.f. Suggest reasons, if any, which might make the above confidence interval inappropriate in a practical situation.
11 -
94 (i) Explain clearly the difference between a poillt estimate and an illterval estimate of a population parameter in the theory of statistical inference. (ii) Suppose that Xl> X2, ••. ,Xn are 11 independent observations from a normal population with mean /.L and variance (J'2. If i and S2 are the sample mean and sample variance respectively, indicate, without proof, how these sample quantities may be used to obtain a confidence interval for /.L. An experimenter plans to make two further independent observations Xn+l> Xn+2 from the above normal population. If A is a known constant (0 < A < 1), prove that the 99 per cent confidence interval for the linear
347
SUPPLEMENT
[unction
~,
X-tOoOIXS
[1-+1-2,\(1-,\) ]~ :S;;L:s;;x+toooIXs[1-+1-2'\(1-'\) J" 2
11
1 1 '
Mre tOoOI is the one per cent point of Student's distribution with 11 -1 d.fo Also, show that, irrespective of the value of ,\ in its permissible range, ,he above confidence interval for L must be greater than the interval W
11 +2)~ x±too01Xs ( 2;- .
9S If
XI' X2,' •• , X are independent observations of a normally distributed rariable X with mean IL and variance u 2 , find the least-squares estimates of ~ and u 2 • Hence indicate, without proof, how an exact test of significance based on Student's distribution may be used to test any null hypothesis about IL· Given the 11 sample observations, an experimenter wishes to obtain another 11 independent observations of X. If Xo is the mean of these 11 observations (yet to be made), prove by a suitable extension of the argument leading to Student's distribution that the 100(1- T)) per cent (0 < T) < 1) confidence interval for the difference x - Xo is Il
where x and S2 are the sample mean and variance of the already observed 11 values of X, and t( T); 11 - 1) is the lOOT) per cent point of Student's distribution with 11 -1 dJ.
96 In 11 Bernoulli trials with a constant probability of success p, w successes were observed. Show that for the probability distribution of the random variable w the third central moment is 1L3(W)=l1pq(q-p),
where p+q=1.
Prove that an unbiased estimate of 1L3( w) is
T
=
113 p*(1- p*)(1-
2p*)
--''-----'------=.----'-
(11 -1)(11 - 2)
,
p* being the observed relative frequency of successes. Also, show that for large 11 var(T)-npq(1-6pqf.
97 In a plant-breeding experiment, the observed frequencies of progeny in the four mutually exclusive classes AI> A 2 , A 3 , A4 were Ill> 112, 113, 114 respectively (I I1j == N). On a genetical hypothesis the corresponding probabilities for the four classes are -&(2+0), -&(2-0), -&(1-0), i6(1+30) respectively, Where e is an unknown parameter. Derive the maximum-likelihood equation for e, and show that the
EXERCISES IN PROBABILITY AND STATISTICS 348 large-sample variance of the estimate 8 is 2 ( A) _ 4(4- ( )(1- 6)(1 + 36) var 6 3N(5 + 26 _ 4(2)
Further, suppose there was some error in the classification of the A A3 progeny, though the Al and A4 plants were classified correctly. PO~I and observed frequencies in the A2 and A3 classes and then derive the equ t~hc for 6*, the maximum-likelihood estimate of 8. Verify that for large sam;I~~1l var( 8*) 4(3 - 28)(5 + 28 - 4(2) var(6) (1-8)(2-8)(29+328)' and hence show that var( 8*) > var( 8). 9d~ . In a Plantf-breeding ebxp~rimdent the observed fdrequencies .of the fOli r
Istmct types 0 progeny 0 tame were al> a2, a3 an a4 respectively whe . On a genetical hypothesis the expected proportions in the fo~c classes are 1(2 - 8), i(1 + 8), i8 and i(1- 8), where 8 is an unknowl~ parameter such that 0 < 8 < 1. Find the equation for 8, the usual maximum. likelihood estimate of 8. Alternatively, a simply calculable linear unbiased estimate 8* may he derived by equating the linear function
L aj == N.
X==al-a2-3a3+3a4
to its expectation. Show that explicitly 8* = (a2+2a3-a4)/N, and hence derive the exact variance of 8*. Also, prove that, irrespective of the true value of 8 1 (*) ~3 -~var8 2N 4N·
99 A multinomial distribution has k (;;a.2) distinct classes and in a random sample of N observations from the distribution, the observed frequencies in the k classes were at> a2, ... , ak respectively. The corresponding expected frequencies were ml> m2, ... , mk (Lk=1 ai = L~=1 mi = N). Assuming that the mi's are all functions of an unknown parameter 8, prove that the equation for 8, the maximum-likelihood estimate of 8, is
t
[ai . dmiJ =0. i=1 mi d8 6=6 Also, show that for the linear function X==t a;.dm i i=1 mi d8
t
l.. (dmi)2. i=1 mi d8 In a particular breeding experiment with a variety of Papaver rhoeas. there were four distinct classes with the expected frequencies var(X) =
E(X)=O;
N
N
N
N
2
4 (3-28+8 2), 4 8(2-8)'4 8(2-8), 4 (1-8).
349
SUPPLEMENT
I'erify that in this case 2N[1 + 2(1- 6)2] var(X)= 6(2-6)[2+(1-6)2]"
100 The serially correlated observations E(Xj) = 0, corr(Xb If X and
Xi+l)
var(Xj) = u = p,
2,
Xl> X2, •••
,x..
are such that
for i = 1, 2, ... ,n;
corr(xb Xi+k) = 0, for k;;:;. 2.
S2 denote the mean and variance of the sample observations, therl
rrove that u2 (i) var(x) = - [1 + 2p(n -l)/n]; n (ii) E(S2) = u 2(1-2p/n); and (iii) -~:s;;p:s;; 1.
Also, if another linear function of the observations is defined as
2
n
xw = n ( n+ 1) 1'=1 L vx", Ihen verify that 2u 2 var(xw ) = 3n(n+ 1) [2n(1 +2p)+(1-4p)].
Hence deduce that, as n ~ 00, the limiting efficiency of x'v as compared with is i.
j
101 If X and Yare two correlated random variables having finite means 3nd variances, define cov(X, Y). (i) Assuming that X and Yare positively correlated and var(X);;:;. var(Y), prove that cov(X, Y):s;;var(X). (ii) Prove that as a first approximation
(Y X) _ cov(X, Y) cov\X' - E(X) . Hence deduce that to the same order of approximation
Y)]
(D E(Y) [ cov(X, E\X)=E(X) 1- E (X)E(Y) . 102 Define the product-moment correlation coefficient between two continuous random variables having finite, non-zero variances, and prove that the coefficient must lie between -1 and + 1. Suppose Xl> X2, ••• ,x.. are random variables such that E(Xj) = p.
x
var(xi) = u 2 , and corr(Xj, Xj) = p, for it- j
II is the mean of the Xi and Yi (Yi =
Xj -
(i = 1, 2, ... , n).
x) are the deviations of the
Xi
from
350
EXERCISES IN PROBABILITY AND STATISTICS
i, prove that (F2 (i) var(i) = - [1 +(n -1)p]; n (11") var (Yi) = (n -1)(1- p)(F2 ,
for i
n
(iii) corr(Yi> Yj) =
-~1' n-
for
i1= j =
=
1,2, , , , , n;
and
1, 2" , " n,
103 If Xl> X 2 " " , XN are correlated random variables with finite m ' and variances, prove that edlJ\
va{~ Xi] = i~ var(Xi)+ i~ j~i cov(Xi, Xj), Suppose that an unbiased coin is tossed n times and the number of tim' the sequence a head followed by a tail (HT) is observed. If Zi is a randoC\ variable which takes the value unity if the sequence HT OCcurs at tI:l~ (i -1)th and ith trials considered together and zero otherwise, express SC the random variable denoting the total number of times HT occurs in the ';; trials, in terms of the Zi' Hence show that (i) E(SII) =i(n -1); and (ii) var(SIt) = 'Mn + 1),
104 (i) If X and Yare two random variables with finite variances, prove that the random variables X + Y and X - Yare correlated unless var(X) = var(Y), (ii) Given that Xl> X2, ••• , XII are serially correlated observations such that and var(xi) = (F2, for i = 1, 2, ... , n, where p and (F2 are parameters (-1 < P < 1 and (F > 0), prove that the variance of i, the average of the Xi is
[1
(F2 + p _ 2p(1- pll)] n I-p n(1-p)2 .
lOS (i) If Xl> X 2 and X3 are uncorrelated random variables, prove that cov(Xl + X 2 , X 2+ X 3 ) = var(X2), (ii) The serially correlated random variables that for any s ~ 1
XI> X2, • , , , XII' , , •
are such
E(xs ) = 0;
\p\
LXv; v=r+l
S3=
L
v=n+l
Xv·
SUPPLEMENT
351
prove that
a2pll-r+l(1_ pr)2
COV(SI,S3)= ~ertce
(l-p)2
deduce that if COV(Sb S2) = COV(Sb S3), then 2pll pr =
1 + pll·
,06 Prove that the product-moment correlation coefficient between any I\VO
random variables with finite non-zero variances must lie between -1
JU d
+1.
The random variables Xl> X2, X3, .•• are uncorrelated and each has zero l1Iean and constant variance a 2 • Given any integer n > 1, another sequence 01 random variables Yt> Y2, Y3, ... is defined by the equations 1 1 11 -
Yi=-
n
L Xi+;,
fori=I,2,3, ....
;=0
II r is another non-negative integer, prove that for any fixed i ifr;?;n
n -r n
if r
107 Prove that the moment-generating function of a Poisson variable X with finite mean p, is
E(e 'X ) = exp p,(e' -1). If Xl> X2> ••• , XII are random observations of X, then prove that the sample mean i has the moment-generating function
exp[np,(e l/ll -1)]. Hence deduce that the cumulant-generating function of the standardized random variable has the series expansion
so that, as n ~ 00, the probability distribution of z tends to normality. 108 Prove that the probability-generating function of a Poisson variable X with mean A is E(8 = exp A(8 -1). X )
If XI and X 2 are two independent Poisson variables with means AI and
"-2 respectively, show that the joint probability-generating function of Xl and ¥=X I +X2 is E(8i". 8n = exp[A 1 (8 1 82 -1)+ A2(82 -1)].
352
EXERCISES IN PROBABILITY AND STATISTICS
Hence prove that the marginal distribution of Y is POisson W'th Al + A2 but that the conditional distribution of Xl given Y == j is 1
e)(l: P)'(l !py-r,
P(XI = r I Y = j) =
Ilh:"u
for 0:S;;X 1 :S;;j,
where P==Al/A2' Discuss briefly how this conditional distribution may be used to t ' equality of the means of Xl and X 2 , CSllhl'
109 A discrete random variable X denotes the number of successe' ' sequence of n Bernoulli trials with probability P for a success in an';~ 1, a Prove that the probability-generating function of X is n.11
t
E(OX) = (q + pO)",
where P + q = 1.
If Xl and X2 are i~depend~nt, random variables respectively denoting tl1l' number of successes m two dIstmct sets of nl and n2 Bernoulli trials w'lI success param.eters PI and P2, prove that the joint probability-general)') 1 , , ( Ill! functIon of Xl and Y = Xl + X 2 IS O( 01> ( 2) == E( O~· , On = (ql + PI 01 (2)n'(q2 + P2 ( 2)"2,
(PI + ql = 1; P2 +q2 == I\.
Hence show by suitable differentiation of 0(01) ( 2 ) that (i) P(Y=r)=
i
v=O
(nl)( ]1
r
~2
v
)Plq'i·-vp2-vq22-r+v,
for 0:S;;Y:S;;n 1 +11 2;
and
(nl)( n2 ) , (ii) P(Xl=sl Y=r)= r s
r-s p
L (n1)( r-]1 n2 )p v
v=O
]1
where p == Plq2/P2ql> and it is assumed that in both (i) and (ii) the binomial coefficients are zero for inadmissible values of r> nl and r> n2' Discuss very briefly how the conditional probability distribution (ii) may be used to provide a test of significance for the null hypothesis H(PI = P2)'
110 A discrete random variable X has the probability distribution defined by the system of equations P(X = r) = Pn
for integral values of X;;. 0,
Define the probability-generating function of X and show how this functioll may be used to evaluate the factorial moments of X. Conversely, if /Ll/) is the jth factorial moment of X about the origin, prove that 1
00
(-I)'
L -,-#L(r+,)' r, ,=0 s,
Pr = --,
Use this result to verify that if, in particular, #Lej) = (n + j _l)(j)pi,
where n, a positive integer, and P (0 < P < 1) are parameters, then Pr= ( n+r-1) r p r(1+p)-
SUPPLEMENT
353 1 In sampling from a binomial population with success parameter k I~pen~ent samples are obtaine?, there bei?g ni observati(;lns in th/ith pie (I == 1,2, ... , k). The relatIve frequencies of successes In the samples ~ pt, p~, ... ,pt. Prove that the linear functions x= .¢
itl
niPf/N and
y=
Jl
(tl
pf/k
ni ;';:N)
both unbiased estimates of p, but that var(x) = pq/N and var(y) = pq/kno (p +q = 1)
here no is the harmonic mean of the ni' Determine the moment-generating functions of the standardized random ,ariables w = (x - P)(N/pq)! and z = (y - p )(kno/pq)~. I
~ence deduce that if the ni are large then, as a first approximation, both w lnd z are unit normal random variables. Also, if the differences between the .1 are small, show that var(x) -1- v(nJ var(y) ii 2 ' «here I
k
kv(ni);';:
i=l
L (ni -
k
iif,
kii;,;:
i=l
L ni'
112 A finite population consists of the integers 1, 2, ... ,N. A random lample of n «N) integers is taken from the population without replacement. If X and Y are the random variables denoting the largest and smallest integers respectively in the sample, then prove that the joint distribution of X and Y is given by the system of equations
where n~X~N, and 1~ Y~X-n+1. Use this distribution to obtain the joint distribution of X and Z = X - Y. Hence show that the marginal distributions of X and Z are given by the systems of equations
C=!)/(~,
for
n~X~N;
P(Z=Z)=(N-Z)(:=~)/(~,
for
n-1~Z~N-1.
P(X=x)=
and
113 The number of motorists in a large city having one, two, three, ... prosecutions over a given period of time is available, but the motorists who did not have a prosecution cannot easily be enumerated owing to the motorist population of the city fluctuating during that period. This gives rise to a discrete distribution in which the zero group is unobserved. An observed distribution of this kind can usually be well represented by a truncated Poisson distribution with parameter 11-, that is, a Poisson distribution in which the probability of the zero class is ignored. Show that for
354
EXERCISES IN PROBABILITY AND STATISTICS
such a modified Poisson distribution the variance is never greater th mean, and that, in fact, the variance is equal to the mean multipliedan thc the probability of a prosecuted motorist having at least two convict' by /l. IOns. 114 A sample of 2 v + 1 independent observations is taken of a r variable X having a probability density function [(x) defined in thcan~I()lll -oo<X <00. Suppose the observations are ordered and let Xl denotr
115 If X is a unit normal variable, prove that the distribution function of X is 1 1 (-1),x 2r +! (x)=-+- I , for x;;:oO. 00
2 .J2;r=o(2r+l)2r .r!
Prove that the moment-generating function of E(e tIX1) =
2 e!t
2
•
IXI
is
(t).
Hence obtain an expansion of the cumulant-generating function and veriry that the first two cumulants of IXI are KI
= (2/7r)!;
K2
= (7r - 2)/7r.
116 A continuous random variable X has the probability distribution with probability density function n(n-l) a"
~-~ x,,-2(a -x),
for
O~X~a,
where a >0 and n, an integer >1, are parameters. If Xl and X2 arc independent observations of X, then show that the probability density function of the distribution of the ratio u = XI/X2 is n(n-l) n-2[ 11 - ( n - 1)] 2(211 -1) . u u,
f
n(n-l) ( ) u-"[n-(n-l)/u], 22n-l
for
or
0
1 u~ ;
~
and
Use this distribution to verify that for (3 > 1 P(u
l~u
r/{3].
~ (3) = 1- 2(211~~){3"-1 [1- (n: 1
117 Two continuous random variables X and Y have a joint probability
SUPPLEMENT Sity function proportional to ~~ e- Y(1-x)y"'(l-x)"'+13,
JlI
355
forO~X~l;
O~Y
d zero otherwise, where a > 0 and (3 > 3 are parameters. prove that for positive numbers rand s «(3) E(xrys) = (3 B(r+ 1, (3 -s)f(a +s+ l)/f(a + 1).
Also, show that the marginal distribution of X has the probability density function (3(l-x)13-l, forO~X~l;
Jod that of Y (3f(a + (3 + 1) y-13- 1r (a + (3 + 1) f(a + 1) Y ,
for 0 ~ Y <00,
where, in standard notation, the incomplete gamma function Y
r
(m)=_l_ fe-Itm-l dt
f(m)
Y
(m >0).
o
118 A point P is selected at random on OA, a fixed radius of a circle with centre 0 and radius a, such that the distance OP = X is a random variable having a distribution with probability density function f(x) = 12x 2(a -x)/a 4 , for O~X~a. Determine Y, the length of the chord drawn through P and perpendicular to OA, and verify that E( Y) = (15'77" - 32)a
10 Also, show that the probability distribution of Y has the probability density function
4!4 [2a -(4a 2- y2)!]y(4a 2 - y2)l, for
O~ Y~2a.
Hence, or otherwise, prove that the probability that the length of the chord (Y) is less than the radius of the circle is
43-24J3
16 119 A random variable X has the Laplace distribution with the probability density function f(x)
1
= 20- e-lxI/CT, for -oo<X <00.
Prove that the characteristic function of X is E(ei'CT ) = (1 + t 2 0- 2 )-I.
356
EXERCISES IN PROBABILITY AND STATISTICS
Given that YI and Y2 are independent observations of a random variabl Y having the Laplace distribution with zero mean and variance two, prove that the characteristic function of e z = (l-tJ.)YI + tJ.Y2 is E(e iIZ ) = [1 + t2(1-tJ.\2Il + t2tJ.2] ,
,+
where tJ. is a positive constant such that O:so; w s Hence deduce that the sampling distribution of z has the probability density function 1 [(1-11.) e-lzI/O-IL)_1I. e-lzlllL] for -oo
120 Given that Xl, X2, ... are observations obtained in a temporal order such that E(xj) = mj
for j ~ 1 and var(xj )
=
a constant,
0'2,
the autocorrelations between the Xj are defined as Pk
1
= P-k = 2
0'
E[(Xj - mj)(Xj+k - mj+k)]
for k ~ 1.
If the Xj satisfy the linear relation Xj+2 = aXj+l + (3Xj + Bj+2'
where a, (3 are parameters such that var(Xj) remains constant for all j and the B'S are independent random variables with zero mean and constant variance, show that (i) Pk+2=apk+I+(3Pk for k~-l; and (ii) var(B) = (1 + (3)[(1- (3)2-a 2]0'2/(1- (3). Also prove that Pk=
(3(1- (3)(A~-I- A~-l) + a(A~ - An (1-(3)(A 1 -A2) fork~O,
where Al and A2 are the distinct roots of the quadratic equation z2-az(3 =0.
121 Explain briefly (without proof) the relation between the normal and
x2
distributions. If X and Yare independent unit normal variables, prove that the random variables U and V defined by the equations y x ,2 U=_l_ e-! dt V = -i e -112dt 2
..n;
J
,
..n;
J
are independent random variables, each having the rectangular distribution in the (0, 1) interval. If U and V are independent rectangular variables defined in the interval (0,1), show that the random variables lW and Z defined by the equations W = (-2 loge U)~ cos 2'ITV,
Z
= (-2Io~
U)lsin 2'ITV
SUPPLEMENT
357
are independent unit normal variables. Also, show how these results may be used to construct from k + 2 (k ;;:.1) independent realizations of a rectangular random variable defined in the (0, 1) interval a function which has the X2 distribution with 2k + 1 d.f.
122 A continuous random variable X has the semi-triangular distribution
with the probability density function 2x/a 2 defined in the range 0 ~ X ~ a. If XI and x" are the smallest and the largest of n independent observations of X respectively, find the joint distribution of Xl and Xn" Use this distribution to show that the joint distribution of U
= Xl + x"
and
v = x" - Xl
bas the probability density function
!n(n-1)a- 21 (uv)"-2(u 2-v 2), for
O~v~a;
v~u~2a-v.
Hence prove that the marginal distribution of v has the probability density function
nv"- 2 ---:----::-[(n -1)(20. - v)"+1 - (n + 1)v 2(2a - V)"-1 + 2V"+1] for 0 ~ v ~ a; 2(n+ 1)0. 211
and that of u (n + 1)a 2n
for
O~u~a
'
and
123 Two related components in an electronic unit are such that the failure of one tends to affect the life-time of the other. If X and Yare continuous random variables denoting the individual life-times of these components, then a possible model for representing their joint distribution has the probability density function
f(x, y) = atf3 2 exp[-132y-(al +13l-13~x], for O~X~ Y = 1310.2 exp[-a2x - (0.1 + 131 - (2)y, for 0 ~ Y ~ X, where the unrestricted range of both X and Y is from 0 to 00, and al> 0.2, 131> 132 are positive parameters such that Prove that the probability that X ~ Y is 0.1/(0.1 + 131). Hence derive the joint distribution of X and Y subject to the condition X ~ Y, and then show that in this case (i) the marginal distribution of X has the probability density function (0.1 + 131) e-<<x,+f3,)x, for 0 ~ X < 00; (ii) the conditional distribution of Y given X density function
= X
and has the probability
358
EXERCISES IN PROBABILITY AND STATISTICS
Also, if it is known that the life-time of the first component is less Ih that of the second and that the life-time of the former is X == x, prove I ~11 the probability that the life-time of the second component is at least equ ~ilt but ,not greater than twice that of the first is a 10
1-e-132X •
124 A sample of 2 v + 1 (v;;. 1) independent observations is taken of . random variable X having a probability d~nsity function f(x) defined in Ih~ range -oo<X <00. Suppose these observatIons are ordered and let Xl denote the smallest, X2v+1 the largest, and X v +l the median observation of th. C sample. Determine the joint distribution of Xl' X v +l and X2v+!' If another independent random sample of 4r (r;;' 1) observations of X . taken, find the probability that, for given Xl> X v +l and X2v+h these ~~ observations are equally distributed in the intervals (-00, Xl), (Xl> x )r (Xv+l> X2v+l) and (X2v+l> (0). Hence prove that the total probability that;il~ second sample observations are equally distributed between the four inlervals determined by the smallest, the largest and the median observations of the first sample is v 2B(2r+ v+ 1, r+ v)B(r+ 1, 3r+2v+ 1)B(r+ 1, r+ 11) (4r+ 3)(3)B(v + 1, v + 1)B(r+ 1, r + 1)B(2r +2, r+ 1)B(3r+ 3, r+ 1) .
p=----~~----------------~--~------~----~---
125 If X is a normal variable with zero mean and variance u 2 , and II y2 is an independent random variable having the X2 distribution with II cU., determine the probability distribution of the ratio Xlu Z=Y'
Y;;.O.
Further, suppose U is a unit normal variable independent of X and Y, and c is a known constant. Then, by considering the expectation over the joint distribution of X and Y, prove that Ex.y[(X +cY)] = PX,y,u(U ~X +cY) = P[Z ~cl.J1 +u 2 ], where <1>(,) is the distribution function of a unit normal variable , and PX,y,u(-) is the probability of the event in brackets under the joint distribution of X, Y and U. Hence, assuming that u is also known, discuss very briefly how Ex,y[(X +cY)] may be evaluated by using the standard statistical tables.
126 If X is a discrete random variable having the Poisson distribution wilh mean p" prove that the moment-generating function of X about the origin is E(etX ) = exp p,(e t -1). Hence verify that the fourth central moment of X is P,4(X) = p, + 3p, 2. If Xl> X2, ••. ,Xn are independent realizations of X, show that the moment-generating function of their sample average x about the origin is E(e tX ) = exp np,(e t/n -1). Hence prove that
T
x
= - [(3x
n
+ 1)n -3]
SUPPLEMENT
359
, an unbiased estimate of I-'-iX) and that 's
var(T)=; [(61-'-+1)2+ 1~I-'-J.
IZ7 A random variable X has a uniform distribution in the interval (0, a). Ux\ and X2 are independent realizations of X, determine the joint probabil~y distribution of ~ence derive the marginal distribution of v and then verify that for any positive constant (3"; O! 1[ (3 2 P(v;;;'(3)=Z
1-;].
Also, prove that 3lvl is an unbiased estimate of a and that the variance of • • 1 2 Ihis estimate IS 2'0! •
128 (i) The probability that an event X will not be realized in the first n trials is u". Prove that the expected number of trials required for the realization of X is 00
L u..
where
Uo
= 1.
r=O
(ii) In a sequence of n Bernoulli trials, the probability of a success (S) is p and that of a failure (F) q (p + q = 1). If Yn is the probability that the pattern FSSS will not be realized in the n trials, prove that Yn
=
t (n -r r) (_ qp
3)(n-rl/3,
r=n/4
3 where, for any integer n;;;' 4, the summation over r is for r = nand those values of n - r which are divisible by 3. Also, show that the expected number of trials required for the realization of the pattern FSSS is 1/qp 3. Verify that for n = 15, the probability that the pattern FSSS will be realized at least once in the sequence of trials is 4 qp 3(3 - 9qp 3+ 5q2 p 6).
129 Explain briefly the concept of the statistical independence of two continuous random variables, and illustrate your answer by indicating (without proof) the behaviour of the joint sampling distribution of the mean and variance of a sample of N (> 1) random observations from a normal population. A random variable X is normally distributed with mean I-'- and variance 0'2. If Xl> X2 are the sample means and sf, s~ the sample variances obtained respectively from two independent sets of n observations of X, prove that the function
is an, unbiased estimate of I-'- and that ( )
var w =
(n + 1)0'2
2n 2
360
EXERCISES IN PROBABILITY AND STATISTICS
130 Explain very briefly, without proof, the relation between the X2 d' distributions. an I· If X~ and X~ are independent random variables each having th 2 distribution with n dJ., determine the joint distribution of the ran~ X variables Rand e, where om X~= R cos2
e and
X~= R sin2
e.
Use this joint distribution to verify that the random variable X2_X2
W=_I __2=cot2e 2XIX2 has the probability distribution with probability density function 1
~
(
1:)' (1 + w1
2)
for -00< W <00.
B 2'2
Hence show that if F has the F distribution with n, n d.f., then
t= In 2
[JF-~] JF
has Student's distribution with n dJ. and, for F> 1, 2t 2 2t [ 1+t2]~ . F=1+-+n In n
[Note: For p>O, r(2p)r(!) = 2 2p - 1r(p)r(p +!).]
131 State briefly (without proof) the relation between the unit normal and X2 distributions. (i) If a continuous random variable X has the Laplace distribution with probability density function f(x) =~ e- 1xl , for -oo<X <00, prove that the characteristic function of X is E(e itX ) = (1 + t2)-I. (ii) If Y1 and Y2 are independent unit normal variables, determine the joint probability distribution of the random variables Ul=~(Yl+Y2)
and
U2=~(YI-Y2)'
(iii) By using the above results, prove that, if Z1> Z2, Z3 and Z4 are
independent unit normal variables, then the random variable
W=ZI Z 2+ Z 3 Z 4 may be expressed as half the difference of two independent X2 ,s each having 2 dJ. Also, show that the probability distribution of W is the same as that of X.
132 Cards are dealt out one by one from a standard pack of 52 cards of which four are aces and the rest non-aces. If X is the random variable denoting the number of cards which precede the appearance of the third
SUPPLEMENT
361 and Y another random variable ~enoting the number of cards which plCn up after the appearance of the thIrd ace and before that of the f h (.t, prove that ourt (.t,
P(X = r) =
(~)C~82)(5:_r) /
er2) , for
2~X~SO,
and P(Y=s I X= r) = (50s-r)(51_1r_J / (51s-r), for
O~ Y~50-r.
Hence derive the unconditional probability distribution of Y. Also, show by considerations of random sampling that E(Y)= ~8. [Note: For
n+p;;;'k+1,
L~=l
(r+p)(k+1)={(n+1+p)(k+2)_(1+p)(k+2)}/
(k+ 2).] 133 A continuous random variable X has the probability density function f(x) defined in the interval -oo<X <00 and the mean of X is fL, an unknown parameter. In principle, the average of n independent realizations of X is an unbiased estimate of fL. However, an occasional sample will, due to the
sampling process, contain one or more observations from the upper tail of the distribution of X. When this occurs, and the sample size is small, the sample average will tend to exceed fL by a considerable amount. A "better" estimate can be obtained by using a procedure which reduces the effect of large sample observations. Suppose a random variable Z is defined such that Z = X if X < t and Z == t otherwise, where t is a specified quantity. Show that E(Z) = PfL, + qt
and var(Z) = p[u~+q(t- fL,)2], where fL, and u~ respectively are the mean and variance of the probability distribution of X truncated on the right at X = t (that is all values of X> t are ignored and the distribution is normalized) and
,
P=
J
q = 1-p.
f(x) dx,
-00
Hence determine the mean and variance of the average of n independent observations of X in which values >t are replaced by t individually. 134 If
Xl> X2, ••• ,
XII are equi-correlated random variables such that var(xJ=u7, fori=1,2, ... ,n;
and corr(Xi> Xj) = P a constant, for i =/= j, evaluate the variance of X, the arithmetic mean of the Xi' Hence deduce that
1
362
EXERCISES IN PROBABII,ITY AND STATISTICS
Also, if Yh Y2, ... 'YII are the deviations of the Xi from show that, for any given i,
L
cov(x, Yi) = -1 [ (1 - p )af + pai II n j=1
aj
x
(Yi == Xi -
x),
] - var(x).
Hence prove that a necessary and sufficient condition for the x. (._ ~, 2, ... , n) to have equal variance is that the Yi should be uncorrelated W'ith
x.
135 Three discrete random variables Xl> X 2 and X3 have the joint probability-generating function E«(JX, (JX2(JX3) = ~ 1 2 3 211
[1 +!!.211 {I + (J2(1211+ (J3)1I}1I]1I
where n is a positive integer. Use this probability-generating function to prove that
+ 1)11-1 2 11 (11+1) ;
n 2(211
.
(I) P(X2 = 1) =
211(11-1) (ii) P(X1 = 11 X 2 = 1) = (211 + 1)11-1 ; and ... [211 (11+1) + (211 + 1)11 ]11 (m) P(X3 = 0) = 2 11 (11 +11+1) 2
Also, if Z
= Xl + X 2 + X 3 , then show that E(Z) = n(n 2 + ;n +4)
2
136 If Xl and X2 are independent observations of a normal variable X having mean /L and variance a 2 , prove that U=X1-X2
and
V=X1+X2
are independent and normally distributed random variables with means 0, 2/L respectively and the same variance 2a 2 • Use the probability distribution of u to verify that E(lui) =
2a .[;'
and hence deduce the expectation of the range of Xl and X2. Further, given three independent observations of the random variable X, verify that the sum of the moduli of the differences of the observations taken in pairs is equal to twice the sample range R. Hence deduce that for a sample of three independent observations of X E(R)=
~.
137 Three players A, Band C agree to playa series of games with a biased coin which has a probability p (0 < P < 1) for a showing a head. In each game, each of the three players tosses the coin once, a success being
SUPPLEMENT
363
recorded if the toss leads to a head. The series is won by the first player to obtain a success in a game in which no other player obtains a success. If just 1\\,0 players obtain successes in the same game they continue to play without lite other player who is deemed to have lost the series. If all three players obtain successes in the same game they all continue to play. Determine the probabilities of the possible outcomes of the first game played, from the point of view of a particular player (say A). Hence deduce lit at Un' the probability that A wins the series on the nth game, satisfies the difference equation Un
= (1-3pq)un_l+ 2p3q2(1-2pq)n-2,
Finally, if cf>(z) =
Lr= 1 urz r is the generating function cf>(z) =
SO
forn~2,
q=l-p.
for u'" verify that
pq2z[1-(1-2p)z] [1-(1-2pq)z][1-(1-3pq)z] ,
that cf>(1) =~. Explain very briefly the significance of this value of cf>(1).
138 Two gamblers A and B agree to play a game of chance with initial capital of £a and £b respectively, the stakes at each trial being £1 on the occurrence or non-occurrence of an event E. If E occurs at a trial then B pays A £1, whereas if E fails to occur at the trial, then A pays B £1. The lrials of the game are independent, and the constant probabilities of the occurrence or non-occurrence of E at a trial are p and q respectively, where p+q = 1. A player wins when the capital of his opponent is exhausted. If Un (0 < n < a + b) is the expected number of further trials required for A to win when his capital is in, prove that Un satisfies the difference equation Un = 1 +pUn+1 +qu,.-I'
Hence show that the expected duration of play for A to win is a q_p
(a+b)(qa_pa)pb (q _ p )(qa+b _ pa+b)
if pi= q,
and ab if p = q. 139 A finite population consists of the N elements Xl> X 2 , Ihe mean and variance of the population are defined as
_
1
X=-
1 S2=_-
N
LX;
and
N -1
Nj=l
••.
,XN , and
N
L (Xj-Xf j=1
respectively. Suppose a random sample of n «N) observations is taken without replacement from the population. The sample mean x is defined as the average of the elements of the population included in the sample, but an alternative formal definition of x is 1 N i=-
L ZjXj,
nj=l
where Z b Z2, ..• ,ZN are indicator random variables associated with the individual elements of the population sampled such that Zj is either 1 or 0 according as X j is or is not included in the selected sample. Use this
364
EXERCISES IN PROBABILITY AND STATISTICS
alternative definition of i to prove that (i) E(i) = X;
and
..) (_) (N - n)S2 (11 var x =
nN
140 A continuous random variable X has the distribution function F( defined in the .ra.nge O:s;;; X.~ a. I~ x! < ~2 <: .. < XN are ordered reaIizatio~~ of X, find the Jomt probabIlIty dIstrIbutIon of Xr and x. (s> r). In particul . if X has a uniform distribution in the interval (0, a), show that probabil~' density function of the joint distribution of u = xrla and v = x.la is I y NI (r-l)! (s-r~I)! (N-s)! u r- 1 (v-u)·-r-l(l-v)N-.,
O:s;;;u:S;;;l;
v;:ou.
Hence deduce the marginal distribution of u and the conditional distribution of v given u. Also, given u and some constant Va such that u < Va < 1, prove that P(V > va) = 1 - BA (N - s + 1, s - r)
in standard notation of the incomplete B-function and where A"" (va- u)/(I-u).
141 Given that X and Yare independent negative exponentially distributed random variables in the range O:S;;; X, Y < 00 and with means II n I and I1n2 respectively, prove that
P(X;::;'Y)=~. nl +n2
Determine the joint probability distribution of X and Y given that X;:o Y. Use this probability distribution to derive the joint probability distribution of U = X - Y and Y, given that X;::;. Y. Hence deduce the probability distribution of U given that X;::;. Y. Also, show that the unconditional distribution of IUI has the probability density function
142 In a quantitative study of the spatial distribution of a plant population in an area, a square region was divided into n quadrats of equal size, and the number of plants observed in the region was s. For given s, the distribution of plants inside the square is such that a plant is equally likely to be found in anyone of the quadrats independently of the other plants. Determine the probability that a specified quadrat contains no plants. Furthermore, if X is a random variable denoting the number of empty quadrats in the square region, prove that E(X) =
and
n( 1-~r
SUPPLEMENT
II
365 Also, obtain approximate expressions for E(X) and var(X) when sand ~rJJ in such a way that sIn = A, a constant.
1.3 A continuous random variable X has a uniform distribution in the
'olerval -a";;; X,,;;; a. If X,t+l is the median of a sample of 2n + 1 independent X n +!. Use this probability distribution to prove that, for any non-negative integer r,
~bservations of X, determine the probability distribution of r aT(2n+2)f{!(r+ 1)} E(Ix,t +1 I ) = 22n + 1r(n + l)r{n +!(r+ 3)} .
Hence, in particular, verify that E(lxn+!i) = (2nn+ 1 )a/22n + 1 •
Also, show that for large n a first approximation gives E(lxn+li) = a/~. [Note: It may be assumed that for large m > 0
r(m + 1) = .J21Tm(mle)"'.]
144 Suppose that Xl> X2, ••• ,Xn are serially correlated observations each with zero mean and variance u 2 , and corr(Xj, Xi+k) = P for all i arid k = 1,2 such that 1,,;;; i ,,;;; i + k ,,;;; n and zero otherwise. If x is the average of the Xi> prove that . u2 [ 2(2n - 3)P] (1) var(x) = -;; 1 + n ; and (ii)
E[t
(Xi -
X)2 ]
=
u 2[ n
-1
2(2nn- 3)p
J.
Hence show that as n ~ 00 p ~ -1.
145 Suppose Xl < X2 < ... < x.. are ordered observations of a continuous random variable X having a negative exponential distribution with the probability density function a- I e-(X- I3)f""
X~(3,
where a and (3 are positive parameters. By using the transformation UI = Ur
n(xi - (3)
= (n -
r+ 1)(x,- X,-l),
2,,;;; r";;; n,
or otherwise, prove that the u's are independent and identically distributed random variables such that, for any j, 2u/a has the X 2 distribution with 2 d.f. Hence determine the probability distribution of the ratio n(n -l)(xl - (3)
Comment very briefly how this result may be used to test a specific null hypothesis about the parameter (3.
366
EXERCISES IN PROBABILITY AND STATISTICS
146 State the additive property of the X2 distribution. Suppose X and ~ a~e two indep,endent r~ndom variables such that 2ex and 2 Y/O are both dlstnbuted as X2 s each with .2 d.f., where 0 is a Positive parameter. If Xl> X 2 , ••• , XII are random observations of X, and Yb Y2 .. those of Y, determine the joint probability distribution of the sum; . , y" n
It
L Xi
U =
and
L Yi·
v=
i=1
i=1
Hence, by considering the transformation
u=w/z
v=wz,
and
or otherwise, determine the joint distribution of wand z. Finally, use this joint distribution to deduce that the marginal distribu_ tion of z has the probability density function 2f(2n) (OZ-I+O-I Z)-2I1 Z -l {f(nW '
and that
for O:s;;z
,
E(z) = (2n ~ 1)0 [f(n -t)/f(n)]2.
147 Suppose Xl is the smallest and XII the largest of n random observations of a random variable X having the probability density function I(x) defined in the range -oo<X <00. Find the joint probability distribution of Xl and x . If another random sample of r observations of X is obtained, determine t1~~ conditional probability, given XI and XII' that these observations all lie outside the range (XI' x,,). Hence show that the total probability of the second sample observations lying outside the range of the first sample is A
P = n(n -1)
JI(Xl) dX J [ f/(X) dx 1
r-
X
2 [
1- f/(X) dx T/(x,,) dX II
= f(n + l)f(r + 2)/f(n + r + 1). Also, if r is small and n large, prove that P=
1)]
.
f(r+2) [ r(r+ nr l-~ approximately.
148 Prove that the probability-generating function of a Poisson variable X with mean A is E(OX) = e A (9-1). If Xl and X 2 are two independent Poisson variables with means A1 and A2 respectively, show that the joint probability-generating function of XI and Y=X1 +X2 is E(O;'!.Oi)=exp[Al(OI02- 1)+A2(02- 1)].
Hence prove that the marginal distribution of Y is Poisson with mean Al + A2 but that the conditional distribution of Xl given Y = j is P(X1 = r I Y= j) = where p=AdA 2 •
e)(l :P)'(l !py-r,
for O:s;;XI :S;;j,
SUPPLEMENT
367
Discuss briefly how this conditional distribution may be used to test the equality of the means of Xl and X 2 • 149 Two continuous random variables X and Y have a joint distribution with the probability density function f(x, y) = n(n -1)a-1I211-lxll-2, defined in the triangular region bounded by the lines x = y, x + Y= a, and ,( == O. Find the marginal distributions of both X and Y. Also show that the conditional distributon of Y for given X = x has the probability density function
fey IX =x) = (a -2x)-1,
for x";; Y";;a -x.
Hence verify that var(Y I X = x) = (2x - a)2/12.
E( Y I X = x) =!a ;
150 If (x) denotes the distribution function of a unit normal variable X, and P(O,,;;X";;x)=I{!(x),
then prove that (x )[1- (x)] = ~ - I{!2(X).
Also, by integrating a series expansion of the integrand of I{!(x) , show that 4 x 2 [ 1- x 2+7x ] I{!2(X) = 271' 90 - ....
3
Hence deduce that approximately 4<1>(x)[1- (x)] = e- 2X2 /'7T[ 1 + 2( 71'3~;)X4].
when x is small.
151 A continuous random variable X has the rectangular distribution in the interval (0, a). Prove that P(X";;!a)=!. If Y is another continuous random variable such that the joint distribution of X and Y has the probability density function
f(x, y) =
(
1
a a-x
)'
for 0,,;; Y,,;; a - X; 0,,;; X ,,;; a,
prove that P(Y";;!a) = pea - X - Y ";;!a) =!(1 + log., 2).
In an experiment a particle Xo with energy Eo is subjected to radiation. There is a probability p that Xo will not split, and a probability 1- p that it will split randomly into two particles Xl and X2 with energies EI and Eo- EI respectively such that EI is uniformly distributed in (0, Eo). In a second experiment the unsplit particle Xo or the particles Xl and X 2 are again subjected to radiation under the same conditions as in the first
368
EXERCISES IN PROBABILITY AND STATISTICS
experiment. Prove that the probability that all the particles result in f the two experiments are less than tEo in energy is g rOIl} t(l- p )[1 +!(1- p )(1 + loge 2)2].
152 If X is a continuous random variable having the Laplace distrib t' . h pro b a b'l' . f unction . Wit I Ity d enslty
U
f(x)=texp-ix-8i,
IOn
for-oo<X
prove that the moment-generating function of X about the origin is E(e tX )
=
e t8 (1- t 2 )-1.
Hence show that, if x is the sample mean of n random observations of X then the standardized random variable '
x-8 z=--
mn
has the moment-generating function (1- t 2 /2n)-n. Further, verify that the coefficients of skewness and kurtosis of the probability distribution of z are 1'1 = 0 and 1'2 = 3/n respectively, and thai for large n, z is approximately a unit normal variable. Alternatively assuming that an approximating non-normal distribution representing th~ behaviour of z has the probability density function proportional to e-!z2(1 + az 2+ (3z4),
determine the normalizing factor of this probability distribution. Also, if this approximating distribution has the same first four cumulants as the exact distribution of z, prove that 6
a = -6{3 = - 8n + 3 .
153 A collection of nl unbiased dice, each of which is coloured red, and another collection of n2 unbiased dice, each of which is coloured white, arc rolled together once. The numbers shown uppermost on the nl red dice total Xl and those on the n2 white dice X2' The collection of white dice is rolled once again and the sum of points obtained is X3' Assuming that each of the nl + n2 dice is numbered from 1 to 6, show that the product-moment correlation between XI + X2 and Xl + X3 is nl/(nl + n2)' Also, determine this correlation when it is given that X2 + X3 = c, a constant. Find the values of these two correlation coefficients when (i) n l = n2; (ii) nl ~ 00 and n2 remains finite; and (iii) n2 ~ 00 and nl remains finite. Explain briefly the meaning of these limiting values.
154 A continuous random variable X has a probability density function proportional to (1 + X 4 )-1 defined in the range -oo<X <00. Determine the proportionality factor and show that the random variable Z =(1 + X4)-1
has the beta distribution of the first kind with parameters ~ and standard notation.
!,
in
SUPPLEMENT
369 If Y is another independent random variable having the same probability J~tribution
as X, prove that the ratio
W=X/Y has the Cauchy distribution with the probability density function
1 7T(1 + w2)' for -00< W <00. Verify that this is also the distribution of W if X and Yare independent unit normal variables. [Note: It may be assumed that f(t)f(l- t) = 7T/sin 7Tt.]
ISS A continuous random variable X has the semi-triangular distribution with the probability density function 2x/ a 2 defined in the range O:s:;; X:s:;; a. If x. is the largest of n random observations of X, determine the probability distribution of XII. Hence show that A
_
a -
(2n+1) 2n x,.
is an unbiased estimate of a with variance a 2 /4n(n + 1). Also, derive the sampling distribution of the standardized random variable
a-a
z=--s.d. (a)
and then prove that, as n ~ 00, this distribution has the limiting form with probability density function
e- l
for -00< z:s:;; 1.
156 A savings bank permits a client to withdraw money not more than once a week. It is observed from the past experience of the bank that the probabilities of a client withdrawing money in the week preceding and the
one following Christmas are PI and P2 respectively, and the probability of a client using the facility in both weeks is P12, where O
E(Oi". 0~2) = [1 + Pl(OI -1)+ P2(02-1)+ pdOI -1)(02 -1)]". Further, assuming that al2 PI2=-, n where at> a2, and al2 are finite constants, determine the limiting form of the above probability-generating function as n ~ 00. Show that for the limiting joint distribution al2 corr(Xt> X 2 ) = ~. vala2 Hence verify that in this limiting case if Xl and X 2 are uncorrelated then they are also independently distributed random variables.
370
EXERCISES IN PROBABILITY AND STATISTICS
157 Two continuous random variables X and Y have the joint distribur with the probability density function IOn 0: - 2 yO< (y ) no: + 1) . (1 + x)13 . exp - 1 + x '
(3 -
defined for O:s;; X, Y < 00, where 0: (> -1) and (3 (>0) are parameters sUch that (3 > 0: + 2. Prove that (i) the marginal distribution of Y has the probability density function «(3 -0: -2)g(y) no: + l)y13-o<-l '
o:s;; Y
where Y
g(y) ==
Je- w 13 w
2
dw;
o
(ii) the conditional distribution of X given Y = y has the probability density function
(Y )
y13- 1 g(y)(l+x)13
exp - - -
l+x '
and
...
(Ill) E(X I Y = y)
e- Yy13- 1
y
= «(3 -2)g(y) + (3 _2-1.
Also, if two new random variables U and V are defined by the relations
1
Y U= l+X
and
V=l+X'
show that U and V are independently distributed and determine their probability distributions explicitly.
158 A continuous random variable X has the exponential distribution with the probability density function 0:- 1
exp -(x-(3)/o:,
for
X~(3,
where 0: and (3 are positive parameters. Suppose that Xl> X2, •.• ,XII are /I independent observations of X. Given that Xl is the smallest of the /I observations, prove that the distribution of Xl has the probability density function no:- l exp -n(xI - (3)/0:, for Xl ~ (3. Hence prove that
Also show that, for given Xl> the conditional distribution of the probability density function o:-lexp-(Xj-XI)/O:,
forxj~xl'
Xj
(if 1) has
SUPPLEMENT
371
Ose this probability distribution to show that
E[~ (xi -
XI)] = (n -l)ex,
and hence verify that
1
L"
ex*=-(xi-Xl) 11 -1 i=2
and
~*=XI
~
respectively.
are unbiased estimates of ex and
159 Given a random sample of n observations Xl' X2,' .. , X" from a unit normal distribution, two statistics t and ware defined as
x~
t=-s-
and
" / Vi~Xf, ,-" W=i~Xi
where x and s are the sample mean and standard deviation respectively. Prove that (i) -JIi ~ w ~ JIi; and (ii) t 2 = (n -1)w 2
n-w 2 Hence deduce that t is a monotonic function of w. Starting with the joint distribution of x and s, use the polar transformation
x~= r sin 6,
sJn -1 = r cos 6
to prove that the probability density function of the distribution of u = wlJli is 1
BH, ~(n -1)}
.
2 !(,,-3)
(1- u )
,
Hence show that the probability distribution of v = Il.Jn -1 has the probability density function 1
2
BH, ~(n-l)}" (1 +v)
_!"
,
-oo
[Note: It may be assumed that the probability density function of the joint distribution of x and s is (n -1)!(,,-J)~
2!("-2)rmf{~(n -1)} exp[-Mni 2+(n -1)s2}]s,,-2.
]
160 In the study of accuracy of shooting at the centre of a two-dimensional target, the horizontal and vertical distances of the point of impact from the centre are known to be two random variables Xl and X 2 respectively having a joint bivariate normal distribution with zero means and second-order central moments uf, u~ and PUIU2 in standard notation. The random variable R defined by
372
EXERCISES IN PROBABILITY AND STATISTICS
is known as the radial error and is taken to be an appropriate measure of t accuracy of marksmanship. Prove that if a rotational transformation is ma~c such that C YI = Xl sin 6 + X 2 cos 6, where _1(-2 Pcr l cr2 ) 6 =It 2 an 2 2' crl-cr2
then the random variables YI and Y2 are independently distributed with zero means and var(YI ) == cr~, = t(cr~ + cr~) +t[(cr~- cr~)2+4(pcrlcr2)2p; var( Y2) == cr~2 = ~(cr~ + cr~) -
mcr~ - cr~)2 + 4(pcr1cr2)2J!.
Hence, or otherwise, show that the moment-generating function about the origin of R 2 is E(e'R2) = [(1- 2tcr~,)(1- 2tcr~2)]-~
and then verify that the jth cumulant of R2 is Kj(R 2 )
= 2j-l(cr~~ + cr~~
. (j -i)!
ANSWERS AND HINTS ON SOLUTIONS Supplement
--------------------------------------------------Finite differences and summation of series
L" r(r+4)[l(r+1)(3)-2r]
=
r=1 n
11
=t L (r+4)(5)-2 L [(r+3)<4)-2(r+1)(2)] n
-2
L [(r+2)(3)+(r+1)<2)-3r] r=\
= h(n + 5)<6) - 2[k(n + 4)(5) -~(n + 2)(3) +!(n + 3)(4) +Hn +2)(3) -~(n + 1)(2)] =-k(n + 1)(2)[5n 4 + 34n 3 -14n 2 -331n -144].
FOrll=1, sum=-10. r
II
L
(ii) Sum =
r(r + 3)
r=1
L
S(2)
.=1
II
L r(r+3) .1(r+1)(3)
=
r=\
=l L"
[(r+3)(5)-2(r+2)(4)-2(r+1)<3)]
r=\
= HHn +4)<6) -hn + 3)(5)-!(n +2)(4)] = -k(n + 2)(4l[5n 2 + 23n + 9].
For
11
= 2,
sum = 20. n-l
(iii) Sum= =
L
r
r(r+2)(r+4)
L (s-1)s(s+2)
r=\
.=2
'1-1
r-l
r=1
1=\
L r(r+2)(r+4) L [(t-2)<3)+(t+1)(2)]
11-1
=L
r(r+2)(r+4)[!(r+2)<4)+Hr+ 1)(3)]
r=\
373
374
EXERCISES IN PROBABILITY AND STATISTICS 11-1
L [(r + 5)(7) -
=!
6(r + 4)(6) + 3(r + 3)(5) + 3(r + 2)(4)]
r=1 11-1
+i L
[(r+4)(6)-3(r+3)(5)-3(r+2)(4)]
r=1
= Uk(n + 5)(8)-~(n + 4)(7) +Hn + 3)(6)+~(n +2)(5)] +Ht(n + 4)(7) -!(n + 3)(6) -~(n + 2)(5)] = 4Ao(n + 2)(5)[15n 3+ 100n 2+ 125n -144]. For n = 3, sum = 384. (iv) Sum=
n
r
r=1
0=1
L (r+1)(r+2)(r+4) L
[(S+1)(3)-4s(2)]
n
=
L (r+1)(r+2)(r+4)[Hr+2)(4)-~(r+1)(3)] r=1 n
L [(r + 5)(7) -
=!
5(r + 4)(6) + 2(r + 3)(5) + 2(r + 2)(4)]
r=1 n
L [(r+4)(6)-2(r+3)(5)-2(r+2)(4)]
-~
r=1
= Ut(n +6)(8)-~(n + 5)(7)+i(n +4)(6)+~(n + 3)(5)] -~[Hn + 5)(7)-i(n +4)(6)-Hn + 3)(5)] = 1O~80(n + 3)(5)[315n 3+ 1005n 2- 4850n -8936]. For n=2, sum =-144.
J: (r+2)(r+6)[~ (S+1)(2)-7:~
(v) Sum =
n+l
=
L
S+8r]
(r+2)(r+6)[i(r+ 1)(3)-~r(2)+8r]
r=1 ,.+1
,.+1
r=1
r=1
=i L [(r+3)(5)+3(r+2)(4)]_~ L
[(r+2)(4)+5(r+1)<3)+5r(2)]
n+l
+8
L
[(r+2)(3)+5(r+ 1)(2)+5r]
r=1
= Hi(n + 5)(6)+~(n +4)(5)]-UHn +4)(5)+Hn + 3)(4)+i(n +2)(J)] + 8[Hn + 4)(4) +i(n +3)(3)+HI1 +2)(2)]
= 12(n +2)(2)[4n 4+ 12n 3-235n 2+411n +6048]. For n = 0, sum = 168. n-2
2 (i) Sum =
L
t2(t + 1)2(t + 5)
1=1 n-2
=
L
[(t+4)(5)-3(t+3)(4)-6(t+2)(3)+6(t+ 1)(2)]
1=1
=i(n + 3)(6)-~(n +2)(5) -~(n + 1)<4)+2n(3) = ron(3)[5n3+ 12n 2-44n +9]. For
11
= 3, sum = 24.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
375
I"
[f(r + 1) - f(r)] r=1 = f(11 + 1) - f(1), where f(r) == 2 r/(r+ 1).
(ii) Sum =
for
It
= 1, sum =1.
(iii) Sum=
'" 1 L -[r(r-1)+4r-6] r=1 r! '"
=
I
1/t!+4
1=0
'"
I
1/t!-6
1=0
'"
L 1/r!
r=1
=6-e. (iv) If Q r denotes the coefficient of xr (r ~ 0), then Aao = 6; A2 a o = 16; A3ao=6; Akao=O for k~4. Hence by Euler's formula the required sum
2(3-6x+ llx 2 -5x 3 ) (1-X)4
for
Ixl < 1.
(i) Sum="f [(r+1)r+(r+1)-1]( n )xr r=O r +1
1" (n) x' L (n -1) xr - - I r x t
(11-2) ,,-I x' + 11 1=0 t r=O
I
,,-2
=
n(n -1)x
=
1 n(I1-1)x(1 + X),,-2+ 11(1 + X),,-I_- [(1 + x)" -1]. x
=
n(2)x 2(1 + X),,-3[(11 + 3)x + 5].
1=1
4 Difference the given terms of the series. Then note that A3u,. = 2, where II, (r~ 1) denotes the rth term of the series. If u,. is a cubic in r then A3 u,. = 2 for all r ~ 1. Then Ur
=
(1 + Ay-I u1
= Ul + =
e~ 1)AUI + (r; 1)A2 U1+
e;
1)A3U1
1(r 3 - 3r2+ 5r + 6).
If A3 ur 1= 2 for all r ~ 1 then u; differs from the above value by a polynomial in r which is zero for r = 1, 2, ... ,7. Hence the possible values of Ur are
t{r 3 - 3r2+ 5r+ 6) + (r-l)(7)P(r) where P(r) denotes a polynomial in r.
376
EXERCISES IN PROBABILITY AND STATISTICS
Assuming that Ll 3 ur = 2 for all r, we have
f
r~1
For
11
Ur
=![{"(11+1)}2_ 11 (1+1)(2"+1)+511(11+1) ] 3 2 2 2 +611 1 = 12"(11 3 -211 2 +511+32).
= 1, sum=3.
5 (i) The rth term of the series can be written as arx r where ar = 2r2
and
x =!,
for r~O.
Observe that Llkao = 0 for k ~ 3. Hence the sum of the series
o
=--+ I-x
2x 4x 2 +---::(l-x)2 (l-x)3
= 12. (ii) Here the rth term of the series is arx r where ar = !r3 and x = t. Also Llkao=O for k~4. Hence the sum is 13. 6
UL (1961). (i) Since Llu" = U,,+1 -
Un
we have U n +l =
=
(1 + Ll)u" (1 + Ll)2U"-1 = ...
= (I+Ll)"ul. (ii) Assume that the result holds for m -1 so that
Hence Ll"'UI =
L
",-I
(-1)'
(m -1)Llu",_ro
where Llu,,,-r = U",+I-r -
r
r=O
Um - r •
The result for m is seen to hold by combination of binomial coefficients n
~
A-I
'- Ur = L.1
(iii)
U,,+1 -
A-I L.1
Ul
r=1
= Ll- 1[(1 =
7
+ Ll)n U1 ]- Ll- t Ut
[(1 + Ll)n -1]Ll- t u 1
UL (1961). Assuming the form of f(x), show that Ll2f(x) = 2h2a2+6h3a3+6h3a3x.
If this is
== a + (3x, then
whence the result.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
8 UL (1962). (i) Put tan-I(x+h)=y and tan- t x=z. Then h
tan(y - z) = -:----:-:-:;-----(x+ h)2- h(x+ h)+ 1 SO that
h ]. h)2 h( x+ x+h)+1 Repeating this argument gives the answer for a 2 tan-I x. II.
L.l
tan - I x = y - z = tan - I [ (
(ii) a sin x = 2 cos(x + h/2) sin h/2, and
a 2 sin x = -4 sin2(h/2) sin(x + It) = -2(I-cos h). (1 +a) sin x, whence the result. (iii) aAl3x = (A13 h -1)A13 and, in general, a rA I3x = (A I3h -1)'A I3x. X
Hence, on summation, (A 13h -1)"-1
II-I
L a Al3x r~O r
=-'------:-:---
Al3h -2
.
Also,
'fl arAl3x = ["f a r] . Al3x = a"a-I-1 . AI3X, r~O
r~O
whence the result. 9 UL (1963). (i) alog(1 + x) = 10g[1 +_h_]
l+x
so that 11.21 L.l
og
(1
+x
)=1
og
[(1+X)(I+X+2h)] (l+x+h?
=IOg[I-(I~Xr{l+ I~J-2] - log [1 - (1 __ (_h l+x
~ x r]
)2 +0(h
3 ).
(ii) a 2 COS x = COS(X +2h) + COS x -2 cos (X + h) = COS(X + h) cos h -sin(x + h) sin h + cos x - 2 cos(x + h) = (cos h - 2) cos (x + h) -sin x sin h cos h + cos X cos 2 h = 2(cos h -1) cos(x + h).
377
378
EXERCISES IN PROBABILITY AND STATISTICS
The result stated now follows by noting that cos(x + It) = (1 + A) cos x 1-cos h = 2 sin2 h/2- h2/2 for small h. and (iii) Ae 2= eX'(e2hx+h2 -1),
and A2ex2 = A[e'(e2hx+h2 -1)] = e'(e4hx+4h2_2e2hx+h2+ 1)_eX2 . 2h 2 . (1 +2X2)+ O(h 3 ).
10 UL (1965). (i) The general result follows on addition and for the particular series note that 1 Ur = f(r) - f(r + 1), where f(r) = 6(3 2)( . r-
3r+1)
(ij) For r = 2 and 3 the recurrence relation gives
28 - 8a + (3 = 0;
-80+28a -8(3
=
0
whence a = (3 = 4. The sum of the series is (1-4x)/(1+2x)2 for Ixl
~-exp(-!n log 0') -1-!n log 0'. 0'
(ii) Here L~'~ 1 Ur
Ur
= f(r) -
= f(1) -
f(r+ 1) where f(r) f(n + 1) ~ i as n ~ 00.
= (2r+ 1)/2r(r+ 1). Hence
(iii) Put x = a-b. Then
g(a,b)=2xa
[1+(1-~rl]+IOg(1-~)
=:: [(!-~)+(!-t)~+(!-t)(~r
+··1
Hence g(a, b»
x3
6a 3
and
whence the result.
12 UL (1975). (i) The expression for f(r -1) - f(r) follows directly. Then the sum of the series to n terms is f(O) - f(n) ~ l/x as n ~ 00. (ii) Set A = [(x -1)/(x + 1)]!. Then log A = ~ [IOg( 1-~) -log( 1 +~) ]
1
1
= --;- 3x 3 + O(x
-5)
.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
379
Hence
1 1 1 -1--+--X 2X2 2x 3 ' (iii) Note that
Ur
13 UL (1973). (i) Note that
Ur
(-xY 1 (-xy+ 1 =--+-. 1 whence r x r+ ~ 1 L Ur = -log(l + x) - - [log(l + x) - x]. r=1 X
= (2 r -l)/(r+ 1)! whence
~
L
r=1
~
ur=t
~
L 2r+1/(r+1)!- L
r=1
r=1
~
=t (ii)
on expansion.
n
1/(r+1)!
~
L 2$/s!- L
$=2
$=2
l/s!,
whence the answer.
(n
r 1 ,,+1 ,,+1 ( L _ )(3- ) =--1 L + 1) 3r, whence the answer on summar=1 r 1 r n + r=1 r tion of the binomial expansion.
(iii) The given expression can be rewritten as x(l + l/x)! 10g(1 + l/x) and the answer is obtained by expansion and retaining terms up to x- 3 • (iv) The given expression = e- 2X '(e2x +e- 2x ) 2x4 ] = 2 e -2X'[1 + 2 x 2 +3+'"
whence the answer on the expansion of e- 2x '.
14 UL (1972). (i) The expression for f(r -1) - f(r) follows on substitution and then S" = f(O) - f(n). (ii) If S denotes the sum of the series then differencing shows that (1- x)S = 1 +2x +(2x)Z+(2x)3+ ... =(1-2x)-1 if 12xl<1. (iii) We have 1 + e = 1 +2x +tx 2+ix 3 + . . " whence, as a first approximation x = te. Now assume that a second approximation is x = te + Ae 2 where A is determined from the equation 1 + e = te + Ae 2 + e~6+M2 Expansion of the exponential and equating the coefficients of gives A = --h.
15 UL (1971). (i) Note that Ur
=
4(r + 3)(3) - 3(r + 3)(2) + (r + 3) + 2 (r+3)!
e2
380
EXERCISES IN PROBABILITY AND STATISTICS
whence the required sum. It now follows that Ur
forX~O
P(X=r)=r(e_1)'
and so
P(X ~ 2) = 1-P(X ~ 1)
= 1 uo+ Ul 4(e -1)
(ii)
L
"'-(3 (
k) = ",-(3 aL coefficient of x(3 in (1 + (3
k=O
X)",-k
k=O ",-(3
=
coefficient of x(3 in
L
(1 + X)",-k
k=O
= coefficient of x(3 in
L'" (1 + x)i j=13
. . (l+x)",+1 (l+x)(3 = coefficIent of x(3 In - - - ------'x x (1 +X)"'+1 = coefficient of x 13 in -x,whence the answer .
(iii)
t (a + kk - 1) ((3 +aa=kk - 1)
k=O
0:>
=
L
coefficient of
Xk
k =0
=
in (1- x)-'" x coefficient of
coefficient of x'" in (1- x)-",-(3,
16 UL (1970). (i) The expression =
t
l+l1a
r
( -1
l+na
)r-l
x 1 +l1a
in (1- xf II
whence the result.
(11)(~)r +na(~)
r=O
X",-k
t
r=1
(n-1) r-1
'
= 0 on summation. (ii) Here u r =f(r)-f(r+1), where f(r) = 4Ir 2(r+ 1f. Hence
L" U = f(l) r
f(n
+ 1).
r=1
(iii) Here f(r+ 1) - f(r) = 14r6 -2r2 whence the expression for r6. Hence f1
fI
L r6=9 L r2+1~[f(n+1)-f(1)], and the answer follows on reduction.
17 UL (1969). (i) Note that (r+s+1)(s+2)-(r+s)(s+2)=(s+2).(r+s)(s+1), whence the
result on summation. In particular, for s = 0, Sr = !r(r+ 1)
and
S,,-r = !(n - r)(n - r + 1).
It now follows that S.5,,-r =;\[(n + 3)(2)(r+ 1)<2) -2(n + 3)(r+ 2)(3) +(r+ 3)<4)].
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
381
Hence II
L SrSII-r =;t[(11 +3)(2)xi(n +2)(3)-2(n+3)Xi(n+3)(4)+!(n+4)(S)] r=1
and the required result follows on reduction. II (11) (ii) Note that r~o r x r+1 =(1+x)II+1_(1+x)lI.
The required result is now obtained by integrating both sides with respect to x from 0 to 1.
18 UL (1968).
(i) The sum of the first n natural numbers is !(n + 1)<2), whence the average.
(ii)
11
II
r=1
r=1
L [r-!(n+1)]2= L [r 2-(n+1)r+;t(n+1)2] (2n + 1)(n + 1)(2) (n + 1)(n + 1)(2) 1( 1)2 6 2 +4 n + n,
whence the result on reduction. (iii) By
definition E( Jr) Hn + 1) - (J2. Hence
Ltl
= L~I=l Jr/n = (J
Jr
and
var( Jr) = E(r) - (J2 =
r
= n 2(J2 =!n 2 (n + 1)- n 2 var(.Jr) =n 2(n+1) 2
[1- l1(n+1) 2V(n) J.
19 UL (1967). (i) Note that ur+l-ur=24r7-8r3, whence 11
n
n
r=1
r=1
L r =1 L r3+2~ L (u r+
1 - U r)
7
r=1
="31 [n(n2+ 1)J2 +2
whence the result.
(ii) Here U r = f(r)- f(r+ 1), where fer) = 1/r2. Therefore II
L ur =f(1)-f(n+1). r=1
(iii) If x takes the values 2r -1 for 1 ~ r ~ 11 and s> r, then the required sum is II
L (2r-1) L (2s-1) r=1
s>r
II =2"1 [{r~1II (2r-1) }2 - r~1 (2r-1)2 J
whence the result on reduction.
382
EXERCISES IN PROBABILITY AND STATISTICS
20 UL (1966). (i) We have Llf(r) =sin[O+(r+ l)ebJ-sin(O+rcp)
= 2 cos[ 0 + eb/2 + reb J sin( eb/2) = 2 sin(eb/2) . sin[O +!( 7r + eb) + reb J. Hence
Ll 2f(r) = 2 sin( eb/2)Ll sinE 0 +!( 7r + eb) + reb J = [2sin(ebf2)f. sin[O+(7r+eb)+ rebJ. The general result now follows by induction. (ii) By definition U r = (1 + Ll)Ur-1 = (1 + Ll)'uo. Therefore
t r=O
(n)(_l),,-r ur = r =
t r=O
(n)(-l),,-r(1 + Ll)'uo r
Lto (;)(-l),,-r(l+LlY].
Uo
For ur =(-lY/(r+1),
Ll"uo=.to (;)<-1)"/(r+1)
t
=(-1)". (n+1) n+1 r=O r+1 (-1)"(2,,+1-1) n+1 ( ...) H 111
_ (r+2)(2)-2(r+2)+ 1
ere
Ur -
=
(r+2)!
[:! -(r: 1)!]- [(r: I)! -(r:2)!]
whence, on summation,
" r [ r~1 u = 1
1 ] [1 1] (n+1)! - 2-(n+2)! .
Probability and statistical theory 21 UL (1965). The total probability of a student entering the second year is 3 5 4 7 4+32·5=8; and the probability of a first-year student getting a degree is
7 20 5 8·21="6·
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT ~e/lce
n)(5)k(1)"-k P(i)= ( k "6 "6 ; P(iii) =
.. -_ L. f P(ll)
383
(n)(5)S(1)"-S --
s=k
S
6
6
st (;)(~)"(~r-s.
zZ UL (1966). Let A be the event that the player has a flush, and B the event that he has at least one spade. (i) P(A) =
(~)(~)
1
(552 ) = 163:60 = 0·0020.
(39)/(52) 5 = 7411 9520 = 0·7785.
(ii) P(B) = 1- 5
(iii) P(BA) = P(B) - P(AB)
= 1-
(39)1 (52) _ (13)1 (52) = 49355488 63441280 5 5 5 5
0.7780.
99 (iv) P(AB) = P(A) - P(AB) = 66640 = 0·0015. 31416 (v) P(A I B) = P(AB)/P(B) = 49 386 904 0·0006. (vi) P(B I A) = P(AB)/P(A) = 1. Z3 UL (1967). P(i)= P(iii) =
1(~
(N:n)/(~
P{ii) = (:)
(:)(~~:)/(~
P{iv) = 1- (N:n)/(~.
(a) Let X be the event that there is at least one bad fuse in the selection of P, and Y the event that all the v fuses selected are bad. Then the required
probability is P(Y I X) = p(YnX)/p(X) = P(Y)/P(X),
whence the result. (b) Let Z be the event that all the· fuses in the second selection of v are good, and Tr the event that there are exactly r bad fuses in the first sample of P, where l:so; r:SO; v. The required probability is P(Z I X) = P(Z n X)/P(X)
where X =
. LT
r=1
Therefore p(znx) =
f P(Z n Tr) f P(Z I Tr)P(T.}. =
r=1
r=1
The required result now follows by noting that
r•
384
EXERCISES IN PROBABILITY AND STATISTICS
24 UL (1968). (i) With replacement sampling the total number of sample points ar" . from the numbers on the three tickets is (611)3. The numbelSlllg sample points leading to the realization of S = 611 is the coefficie~t or 0 611 in the expansion of or H(O)
=
(1- ( 611 )3/(1- 8?
The required coefficient is 18n 2 +9n-2=(3n+2)(6n-1) whe the answer. ' nee (ii) The probability-generating function of S is H(0)/216n 3 and E(S) H'(1)/21611 3 • When sampling is without replacement then the total number of sampl . points is 611(611 -1)(6n -2), and the number of sample points leading to th~' realization of S=611 is also less than 18n 2 +9n-2. From this we have te exclude the sample points corresponding to the integral solutions of x + y :' z = 611 whe~e (a) x = y = z and (b) y = z1= x so that?' +2y = 6n. The solution correspondmg to (a) IS x = Y = z = 2n and the solutions corresponding to (h) are obtained for 1:so; y:so; 3n. The solutions for (b) are 3n -1 in number and hence the total number of sample points are 3(3n -1). Therefore the excluded sample points corresponding to (a) and (b) are 9n -2, whence the answer. 0
25 UL (1969). (i) Suppose Xi is the event that the ith ball drawn is white, i == 1,2. Then . n(n+r) P(I) = P(X1n X 2 ) = P(X2 1 X\)P(X1) = ( ) ( . m +11 m+l1+r)
(ii) Similarly or by symmetry P(ii)= (iii) Suppose
m(m+r) (m+n)(m+l1+r)
Yi is the event that the ith ball drawn is black,
i == 1,2.
Then P(iii) = P(Y2 1X1)P(X,)+ P(X2 1 Y1)P(Y,) 2mn (m + n)(m + n + r) . (iv) P(iv) = P(X2 ) = P(X2 1X,)P(X1) + P(X2 1 Y1)P( Y1) n =-m +11 (v) P(v) = P(Y1 I X 2 ) = P(Y1n X 2 )/P(X2 ) = P(X2 1 Y1)P( Y 1)/P(X2 ) m
m+n+r
Expectation of white balls drawn is 2nl(m + n).
26 UL (1970). Suppose X is ·the event that all r balls are of the same colour, Y tl!l' event that all are black and Z the event that all are white. Then X == Y+Z.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
"here in the first experiment P(Y) = P(Z) =
(J /
385
(2:).
/'he required probability is P(Z I X) = P(Z n X)/P(X) = P(Z)/P(X) =!.
In the second experiment
/'he required probability is
(J
1
Ignoring terms involving N- 2 , this probability is approximately =
[1 + 1- k{2r+ (k -1)}/2~-1 1-k(k-1)/2N - J
=~[1_k(~~1)][1 k(k;~-1)r1
-~ (1+2k~). 17 UL (1979). P(Holmes dies) = xy + (1- x )(1- y) = a P(Holmes escapes) = (1- x)y = a P(Draw) = x (1- y) = 1 - 2a.
(1) (2)
(3)
From (2) and (3), x = y + 1- 3a whence on elimination of x from (1) we obtain the stated quadratic in y. The condition that the quadratic has real roots gives for a> 0, a ;;a;. §. The inequality now follows since from (3) 1- 2a;;a;. o. 28 UL (1971). (i) By the hypergeometric distribution
and
The recurrence relation follows by taking the ratio P(r)/P(r-1). (ii) Let Xk be the event that A has k court cards in his hand (o:s;; k :s;; 13) and Y the event that the game ends in a draw. Then 13
P(Y) =
L P(Y IXk)P(Xk), k=O
386
EXERCISES IN PROBABILITY AND STATISTICS
where P(Xk ) = P(k)
as in (i)
and P(YIXk )=
C6;k)(:~~~)/G~),
=0, for
for
0~k~8
9~k~16.
If it is known that A has all court cards, then the probability of B hav· no court cards is Ing
The percentage error in the approximation is
2 200 7 703x100=~ 200 7· 703
29 UL (1972). The probability that at least one six occurs for the first time on the rth throw of the k dice is
For given k the expected reward is (in £)
Ek =
f 21_kr(~)k(r-1)[1_ (~)k] r=1
=
6
6
1 1-(i)k 1 2 k- 1 • 1-U2)k <2 k- 1 •
Hence, on the average, the player can never be on the winning side if the non-returnable stake to play the game is £2 1 - k • We have
Ek 1-(i)k 1-(fi)k+1 E k+1 =2 ·1-(i)k+1· 1-(f2)k 1-(i)k >2· 1 _(i)k+1 _12 [ (i){1-(i)k-1}] -11 1 + 1_(~)k+1 12 >11> 1, whence Ek > E k + 1 • UL (1973). Suppose Xr (0 ~ r ~ n) is the event that there are r white and n - r black balls in the urn, and Y the event that in the sample of k balls drawn, there is
30
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
387
oply one white ball. Then Il-k+l
P(Y) =
L
P(Y I Xr)P(Xr),
P(Xr) = (;)prqll-r
apd P(YI Xr)=
G)(:~~)/(;)
for
1~r~n-k+1
=0 for r=O.
Ip the second case P(Xo) =0; 31 UL (1975). The probability that the player gets far at the rth trial is (~~y-l. ~~. Hence the expected gain of the player is
r~l 00
(25)r-l 11 36
11a
. 36' a r = 36-25a .
The player, on the average, will be on the winning side if the expected gain >! which gives a >~. 32 UL (1977). Irrespective of the suit of the cards, the following ten realizations of five cards lead to a run: (A, K, Q, J, 10), «K, Q, J, 10,9), ... , (5,4,3,2, A). Since there are four suits and in all 52 cards, the probability of a run is
and that of a running flush is
whence the probability that the player has a run but not a running flush is 40. 255Wn = 81> say. If it is known that the player has the jack of hearts, then the three corresponding probabilities are
The required percentage change is
The initial probability of a run is 128 51.49.13
PI'
388
EXERCISES IN PROBABILITY AND STATISTICS
say and the initial probability of a running flush is
1 51.49.26 = P2, say. When the player has the jack of hearts, these probabilities are chan d ~~PI and ~P2 respectively. Hence the percentage change is ge to
[ ~(Pl- P2)
1J x 100 = 4.
P1-P2
32a
UL (1966). (i) Let X be the event that the examinee knows the right answer and Y
the event that he gives the correct response. Then P(X)=p; Hence
P(X)= 1-p;
P( Y \ X) = 1 - a ;
' P(Y \ X) '= 1/11.
P( Y) = P( Y \ X)P(X) + P(Y \ X)P(X) n(1-a)p+ 1-p n
Then P(X \ Y) = P(Y \ x)P(x)/P(Y) np(1-a) ~1 np(l-a)+ 1-p
as n ~OO.
Also P(X \ Y) = P(X n Y) = 1- P(X) - P( Y) + P(X n Y) P(Y) 1-P(Y) 1-P(X)-P(Y)+ P(X \ Y)P(Y) 1-P(Y) (n-1)(1-p) ~ 1-p <1 npa +(n-1)(1-p) ap +l-p
as n ~OO. (ii) We now have P(Y IX) = 1/(n - m) instead of lIn. Hence
P(X \ Y) = (
(n-m)p(l-a) ) (1 ) 1 ~1 n-11l p -a + -p
.
as n ~ 00 and m fimte
and P(X\ Y)=
(n-m-1)(1-p) ~ 1-p <1 (n-l1l)pa +(n-I1l-1)(1-p) ap + 1-p
as n ~ 00 and m finite. 33
UL (1967). Let A denote the proposal "decrease duty on cigars and tobacco" and B the proposal "increase duty on cigarettes". _ If N(X) denotes the number of M.P.'s voting for a proposal X and N(X) those against X, then we set N(A) =x;
N(B)
= y;
N(A nB)=z;
N(A
nB)=w.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
389
f\lso assu~e that there are 11 M.P.'s in all. Then the conditions of the problem give x 3 y
,l\lso,
I1-X=2:'I1-Y
(1)
z + w + 135 = y w+z-165=I1-Y z-1l0=2(w+llO)
(2) (3)
(4)
N(A nB) =N(A)-N(A nB) = x- z N(A
n B) = N(B) -
N(A
n B) = y- z.
Hence, using the fact that A n B, A n B, A n B and A n B are mutually exclusive and exhaustive events, we have z + (y - z) + (x - z) + w = 11
(5)
Elimination of x, y, wand z from (1)-(5) gives the quadratic in 11 as 11 2-97511+67500=0 with roots n = 900 or 75. For 11 = 900 x=675;
y=600;
N(A
n B) = 420; n B) = 255;
Thus N(A
z=420;
w=45.
N(A nB) =45 N(A
n B) = 180.
34 UL (1968). There are 25 (x, y) sample points. Divisibility by 5 of x 2 + y2 is ensured for the following 9 points.
(1,2), (1, 3), (2,1), (2, 4), (3,1), (3,4), (4, 2), (4, 3), (5,5). Hence the stated probability. Divisiblity by 5 of 2x 2+ 3y2 is ensured for the following 9 points: (1,1), (1,4), (2, 2), (2, 3), (3, 2), (3, 3), (4,1), (4, 4), (5, 5), whence the probability. The point (5, S) is the only one common to the two sets. When 1:5; x, y:5; 5N, then the sample points giving divisibility of x 2 + y2 by 5 are for x = 1, y =2+5r; 3+Sr 0:5;r:5;N-1 for x =2, y=l+Sr; 1+4r for x =3, y = 1+5r; 1+4r for x =4, y =2+5r; 3+5r for x =S, y =5+5r. This pattern repeats itself N times as x varies from 1 to 5N. Hence there are 9N2 sample points out of a total of 25N2 which now give divisibility of x2 + y2 by 5. The probability is independent of N. A similar argument applies for the divisibility of 2X2+3y2 by 5. There are ~ common sample points (S + Sr, 5 + Sr) which give the divisibility of both X2+y2 and 2X2+3y2 by 5.
390
EXERCISES IN PROBABILITY AND STATISTICS
35
UL (1969). Suppose Hr is the event that the rth urn was selected and X the e that the two balls drawn are white. Then vent k
P(X) =
L P(X I Hr)P(Hr)
where P(Hr) = 11k.
r=l
n-r)2 (i) Here P(X I Hr) = (- for 1 ~r~k-l m+n =1 for r = k. Hence P(X) =~
[en ~nr{ n 2(k-l)- nk(k-l)+ k(k -1~(2k-l)}+ 1] for l~r~k-l
(ii) Here P(XIHr)= (n-r)(n-r-l) (m+ n)(m+ n -1) =1 Hence
for r= k.
P(Hk I X)
(n+m)(n+m-l) 2 (k -1)[n - nk +!k(2k -1)]- n(k -1) +4k(k -1) + (n + m)(n + 111 -I) .
36 UL (1966). Let X be the event that the two consecutive throws of the die selectcd show a double-six. Then 6
P(X) =
L P(X I D )P(D j
j)
j=l
where P(Dj ) =!, the probability of selecting D j and P(X I D j ) = 2 for i = 1, 2, 3;
m
=0 =(4)2
for for
i = 4, 5;
i=6
Hence P(X) = is and P(D6 1 X) = P(X I D 6 )P(D6 )/P(X) =~
If D3 is a biased die with probability x of it showing a four or a five, thcn the probability of it showing a six is 2x. Hence 4x = 4 or x = k, and P(X I D 3)= (~f. We now obtain P(X) = ii4 and P(D6 1 X) = ~~.
37
UL (1969). n
G(Ol> O2 ) =
nr
L L O;:O~P(XI = r, X 2 = s) r=Os=O nr
It
=
L Orp(X = r) L 0~P(X2 = s IXl = r) 1
r=O
=
s=O
L 0;: (n) prqlt-r L It
r=O
Itr
r
s=O
= [q + pOI(q + p 02t]lt.
O~ (nr) pSq'''-S
s
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
391
(i) The probability-generating function of the marginal distribution of X 2 is G(l, ( 2) = [q + p(q + p( 2)"]"
whence P(X2> 0) = 1- P(X2 = 0) = 1- G(1, 0). (ii) and (iii) If Y = Xl + X 2 , the probability-generating function of Y is G(O) = [q + pO(q + pO)"]
whence, on differentiation, E(Y) = G'(l)
and var(Y) = G"(1) + G ' (1)[l- G ' (l)].
38 UL (1969). Let X", Y,,, Z.. be the events that after n moves the system is in an A-type, B-type or C-type cell. Then for n ~ 1 X,,-l + Y,,-l + Z,,-l = n, the sample space. Hence P(X,,) = a,,_IP(X.. I X.. - I) + ~"-IP(X" I Y,,-l) + 'Y.. -IP(X" I Z.. _I) and similar equations for P(Y,.) and P(Z..). Symmetry considerations now show that P(X" I X,,-l) = 0, P(X" I Y,.-I) =!, P(X,. I Z .. -l) =i,
P(Y,.I X,,-l) =i, P(Y,.I Y,,-l) =t, P(Y,. I Z,,-l) =!,
P(Z" I X .. - I ) =t P(Z,. I Y,.-l) =t P(Z" I Z,,-l) =i.
Hence the difference equations a" = !~"-I +i'Y"-1 ~Il = iall-l +t~Il-1 +!'YIl-I 'Y" =~a"-l +t~"-l +i'Y,,-1 with the conditions a,,+~,,+'YIl=a"-I+~"-I+'YIl-l=1. Elimination of an and 'Yn gives the following non-homogeneous equation 70= 120~" +27~"_1. The transformation ~" = x,. +¥t reduces this to 40x,. +9x,,_1 =0.
Hence 10
(9)11
~Il = 21 + k - 40
for n ~ 1
where k is arbitrary. For n = 0, ao = 'Yo = t ~o =! and so ~l = flo. Hence using earlier equations,
k = -li9. Therefore,
all =!-*~.. and 'Yll =!-a~". By symmetry, the probabilities for specific A-type, B-type and C-type cells are aa", i~" and h". As n ~ 00, these probabilities tend to i4, i.., /4.
392
EXERCISES IN PROBABILITY AND STATISTICS
39 UL (1967). (i) Let Xo be the initial probability of A winning the series; this i' . I. the probability of A to win whenever the scores of the two PIS d So 'h ' 'h are Ieve,I If XII IS t e pro b a b'l' I Ity 0 f A winning t e series when ayers he .. n games ahead of B (-4~n~4) then the forward diffe" IS , , lenee equation IS for-4
XII=PXII+I+qxlI_l+rx",
x4=1,
X-4==O,
whence PX,t+2-(P+q)XII +1 +qxlI
= O.
The roots of the auxiliary equation are 1 and q/p and so XII =
IX
+ {3(q/p )11,
where the arbitrary constants IX and (3 are determined by using the conditions X4 = 1 and X-4 = 0, Hence p8_ p4-lI q ll+4 XII =
and the stated result for
Xo
8
P -q
8
follows by putting
n =
0,
(ii) This probability is obtained by putting -n for n and interchanging II and q in the expression for XII'
40 UL (1967), The probability that all the k bottles of the first customer are fresh is
Let Xr be the event that r fresh and v - r a day old bottles are sold oul before the selection of the second customer, and Y be the event that she selects k fresh bottles, then v
P(Y) =
L P(Y I Xr)P(Xr) r=O
where
P(Xr) = P(Y I Xr) =
(;)C':') j (m: n) (n k- r)j (m +kn-v),
for r = 0, 1,2, , , , , v.
The stated result is obtained by recombining the binomial coefficients. Again, if Z denotes the event that the second customer gets no fresh milk, then v
P(Z) =
L P(Z I Xr)P(Xr) r=O
where P(Z I X r )=
(m +;-v)j (111 +;-v).
Finally, let W be the event that at least one of the k bottles of the
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
393
second customer is fresh. Then P(YI W)=p(ynW)/p(w) P(Y)
1 - P(Z)' whence the stated result.
41 UL (1966). The difference equation for Un is Un = p2 . U,.-3 + pq . U,.-2 + q . u,.-\> for n ;a. 3 with Uo = 1, Ul = q, U2 = pq + q2 = q. The roots of the auxiliary equation are 1, (-!p)(1 ± i.J3) whence the general solution Un =A+B[(-p)n(cos
~71" +i sin n371")]+c[(-p)"(cos n371" -j sin n371")]
where A, Band C are arbitrary constants. To determine the constants use the initial conditions. This gives B=-JL (l+p)iv'3-q. 2.fj· 1 + P + p2 '
A== _ _ I_
1 + p + p2
C = _JL (1 + p)j.fj+q 2.fj . 1 + P + p2
The stated result follows on reduction.
42 UL (1966). The total number of samples of size n is (~). In any sample of size n, x will be the largest integer if the other n - 1 integers in the sample are drawn from the x-I integers less than x. Hence the probability distribution of X. The assumed equality may be proved as folIows.
xt (=:~)= xt C~~) =
Nf' (t+n+k) 1=0
t
N-n
=
L
coefficient of
ZO
in z-l(1 + z)'+n+k
1=0
= coefficient of
N-n
ZO
in (1 + z)n+k
L {(1 + Z)/Z}I
1=0
= coefficient of ZO in (1+z)n+k[(1+z)N-n+l. z-(N-n)_z] = coefficient of zN-n in (1 + Z)N+k+l
= (N+k+l)= (N+k+l). N-n n+k+l For k = -1, we obtain the sum of the probabilities P(X = x). The expression for E[(X + r -1)<')] is also obtained by the use of the equality whence the required values of E(T) and var(T).
43 UL (1966). The difference equation for Sn is Sn+l =(1 +p)Sn +nR +(I-p)R +p. 2R, and So= R. To solve this difference equation, make the transformation Sn = Xn + a + (3n,
394
EXERCISES IN PROBABILITY AND STATISTICS
where a and (3 are so chosen that the equation is transformed into x,,+1-(1+p)x,. =0 It then follows that
a=-
R(2+p) P
This leads to the stated result for 5" on reduction. To obtain the inequality, observe that R
5" =2"[(1 + p)"{1 +(1 + p)p +p2}-{1 +(1 +p)p+ np}] p
and since 0
<;
(1+ p )n+2=RC;Py(1 +p)"- RC;PYenp
The inequality for n follows on taking logarithms.
44 Suppose there are x shillings (2x half-crowns, 3x halfpennies), y two-shilling pieces and z pennies. Then 5x + y = ~z. If x = 2, then z = 8. Hence, working in halfpenny units, the equation for y is 6 X 1 + 8 X 2 + 2 X 24 + 48 y + 4 X 60 = 406 whence y =2. For equal division of money the probability is
(~)(!)(~)(D(~) / (~~) = ~10909 -0,0476. If x is unknown, then z =2X2; 5x+y =3x 2. Hence, in terms of halfpenny units, the equation for x is 3x X 1 +2x 2 x2+24x+48(3x 2 -5x)+2x x60 =406
which reduces to (x - 2)(148x + 203) = O. Therefore x = 2.
4S UL (1970). (i) Make the transformation z = !(t -1). Then the given integral is 1 (X- l112 1 2 1 (X- l)/2 I 2 1 (-X- l)/2 _h2 dz e J 2Z J2; e- dz = J2; e-'Z dz - J2; .
J
J
(-x-1)/2
whence the result in terms of ct>(.). (ii) There are 2N + 1- n members who are not in the pressure group. If r
is a random variable denoting the number of these members who vote with the pressure group, then r is a binomial variable such that E(r) = !(2N + 1- n); var(r) =i(2N + 1- n). The required probability is P(r~N+ 1-n) = 1-P(r:!S;N-n) =1-
L
N-"
r=O
-ct>(-
(2N + 1- n)(1)2N+l-" r 2
n
)
\.J2N+1-n '
by using the normal approximation for the binomial distribution.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
395
For fixed N this probability is an increasing function of n and if we assume that <1>(3) -1, then P(r ~ N + 1 - n) - 1 if n2~9(2N +1- n),
whence the inequality for n.
46 UL (1967). P(X = n) = Probability that there are w -1 marked animals in the first n -1 members of the second sample x Probability of selecting a marked animal on the nth occasion from a total of N - n + 1 animals of which W - w + 1 are marked
. N-n+1 '
which reduces to the stated result on recombining the binomial coefficients. On reduction (N - n)(N - W) N(N- W-n+w)
g(N) g(N-1)
=
(1- ~ / (1- ~=:).
The stated result follows by considering g(N)/g(N -1) > l. · f · (W + 1)n 1. · d An unblase estImate 0 N IS W
47 UL (1965). The probability that the player loses for the first time at the (r+ 1)th trial is prq and his net gain is rx - a. Hence QO
E(S) =
L (rx-a)prq r=O QO
= -a + pqx L
rpr-1
r=1
d = -a +pqx. dp [(1-p)-1],
whence result.
In the same way, QO
E(S2)=X2q
L {r(r-1)+r}pr-2axp/q+a 2 r=O
Hence var(S). The inequality E(S) ~ a leads to x> a/p since 2q < l.
48 UL (1975). P(r, X = x) =
(:}n1- 6)x-r . (1- p )px-n
396
EXERCISES IN PROBAB ILITY AND STATISTICS
and so
L'"
= (1- p)e
Xpx-n
x=n
Lx
(
r=1
x-
1) or-l(l- oy-r
r-1
'"
L
=(1-p)O
Xpx-n
x=n
'"
=(l-p)6
L (n+t)p'
,=0
'"
=nO+(l-p)O
L tp',
,=0
whence the result on summation.
49 UL (1972). When the housewife has x different coupons, the probability of her obtaining a new coupon is 1- x/n and that of a duplicate COupon is X/II. Hence P(r I x) and E(r I x) =
(1-~) n
n n-x
= --
f
r=1
r(~)r-l n
. on summation.
Hence the expected total number of weeks is "-1
11-1
x=1
,=1
L n!(n-x)=n L t-
50
1•
UL (1971). ~
(i)
L
~
r(k)p(X = r) = f.L k
L e-lLf.Lr-k/(r- k)!
r=a ~-k
=f.Lk
L
e-ILf.LS/s!,
whence the result.
s=Ot-k
(ii) If Y denotes the number of rolls of film sold, then
P(X= r) P(Y= r)= 1-P(X=0)' for Y::::.1. Therefore the probability distribution of Z is P(Z = 0) = P( Y = 1) P(Z=r)=P(Y=r), for 2.:;;Y.:;;5 P(Z = 6) = P(Y::::. 6). Hence 5
rP(X=r)
'"
P(X=r)
E(Z) = r~21-P(X=0) +6 r~61-P(X=0)
1 _ [f.LP(1 ':;;X':;;4)+6{1- P(O':;; X.:;; 5)}], 1-e jJ.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
397
",hence the result. Similarly, 1 E(Z(2) = --_- [/-L 2p(0 ~ X 1-e IL
~ 3) + 30{1- P(O ~ X ~ 5)}]
",hence var(Z). The average return per roll of film sold is (in pence)
42- 6(1-a)+/-Lb. 1-e- 1L
Sl UL (1970). Expected loss =
I e- IL /-L:r. . ar(r+ 1) e-/3r
r=O
-ar e IL =a e- r~1 (r _1)! /-L,[(r-1)+2] 00
=a e- IL e- 2/3/-L2 [
-/3(r-2) r-2 -/3(r-1) r-l] /-L +2e-/3/-L L e /-L , r=2 (r-2)! r=1 (r-1)! <X>
<X>
Le
whence the answer on summation. Since 1- /-L < 0 and /-L e-/3 < 1, we have Expected loss < 3a e 1 -
1L
< 3a.
S2 UL (1969). Assume a Poisson model for the distribution of errors with mean /-L. Then the expected cost of correcting the errors made on a stencil is
I e-IL/-Lr 2r(3r+2) r=O r! . r+1 =2e- I /-L'[3r(r+1)-(r+1)-1] =
1L
(r+1)! r r 1 r+l] 3 L _/-L__ L ~+- L _/-L_ r=l(r-1)! r=or! /-Lr=o(r+1)!'
r=O
=2e- 1L
<X>
[
<X>
whence the result on summation. The residual profit is a -2[3/-L -1
+;
<X>
(l-e- IL )]
and this will be equal to '\a if
(1-'\)a+2 2
3 /-L+1 (1-e-) IL /-L =1+~/-L+i/-L2- ... ,
----=
whence the answer by retaining only the terms in /-L.
53 UL (1968). Suppose Ny calendars are ordered. For 0 ~ x ~ y, the profit is PI = Nx x 30a + N(y -x) x 30~ -Ny x30
398
EXERCISES IN PROBABILITY AND STATISTICS
and for x > y the profit is P2=Ny x30a-Ny x 30
Hence y
00
I
G(y)=30N
L
[(a-{3)x-(1-{3)y]P(x)+30N
X~O
(a-1)yP(x)
x~y+,l
whence the result on using the relation y
00
L
P(x) = 1-
x~y+l
I
P(x).
x~o
Direct substitution shows that G(y + 1) - G(y) = 30N[a -1- (a - (3)F(y)]. Maximum profit is obtained for the smallest value of integer y which makes aG(y)
E(Y2) =
a
f
+ N '\x
e-A(x-l)
dx
1 00
f
=a+N '\(u+1)e- AU du, o
whence the result. Therefore the expected population at the start of the third year is a + (,\ : 1) [a +
N('\ : 1) ]
=
a[ 1 + (,\ : 1) ] +
N('\ : 1
r
Hence the expected population at the start of the (k + 1)th year is a[ 1+ (,\ :1)+ (,\
:1y
+ ... + (,\ :1r-l]+N('\ :1r =
The equation 2(,\ + 1)k
a
[(,\
+ 1)k _,\ ,\ k-l
k] + N (~)k ,\ .
= (/L + 1),\ k follows directly. Then
10g(/L; 1) = -k 10g[1- ,\ ~ 1] ~ ,\ ~ 1 for'\ large. 55
UL (1965). 00
E(T)=
~f x(1+{3x)exp{-~-{3(a-X)}dx
v27T{3 =
-00
2{3
{1
}
-0<11+111 3 fOO a ~2 x(1 + (3x) exp - - (x - {32f dx. 27T{3 2{3
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
399
The integration is completed by using the substitution z = x - (32 a d th . 0 f th i 'Integra.I n en the propertIes e norma Differentiate E(T) logarithmically with respect to a to maximize E(T).
S6 UL (1966).
For small a, (1 - 1~O) 100 ~ e -a. Hence the probability that out of ten
consignments exactly w will be rejected is
Therefore the equation for determining a is
To solve this put z = 1-e-a , where z is small. The terms containing z and Z2 vanish and we have approximately 1-120z 3 ~ 0·985 whence a
=
or z
= 0'05,
0·05.
57 UL (1966). Expected amount payable to proof-reader is 00
T=a+e->-a.P(x=O)+
I
pr-1
a .p(x=r)
r=l
whence the result. O<e->-
and
O<e->-(l-e->-):S;;~.
Therefore
whence the inequality by using p e->- :S;;!. 58 UL (1967). The probability of finding < n defectives in the sample of N is
This is also the probability of accepting the batch after inspection, which now contains only M - N components. For each component sold, the expected net income is 1 - a6. Therefore g(n, 6) = (M - N)(l- (6)
~t~ (~6X(1- 6)N-X
400
EXERCISES IN PROBABILITY AND STATISTICS
whence 1
J
g(n) = E0[g(n, 0)] = (M - N) (1- aO)
~~: (~OX(l- O)N-X . 66(1- 0) dO
o 1
= 6(M - N)
=6(M-N)
"of (N\ J(1- aO)Ox+1(1- 0)N-x+1 dO
X~O x)
o
:~: (~[B(X+2, N-x+2)-aB(x+3, N-x+2)J
6(M-N) .. -1
L
)(4) (x + l)(N - x + l)(N + 4 - ax - 2a), on reduction N+4 X~O 6(M-N) .. = (N+4)(4) Z~1 z(N-z+2)(N+4-a-az),
= (
whence the result on summation over z. For ~ = n/ N, we have g(n)-
(M-N)e 2 [6-4(1+a)~+3ae].
It now follows that g(n) is maximized for
a~ =
1.
59 UL (1968). The expected cost of a packet before devaluation is 00
a
+ {3p, -
~
(TV27T
J(X 2_(T2) exp[-(x -p,)2/2(T2_'Y(X-p,)/(T] dx _00
The stated result now follows by putting u = z + l' and then integrating over u by using .the properties of the normal integral. The expected cost after devaluation is a' + {3' p,' - e!"Y'( p,' - 1'17')2
whence, under the given conditions, , a+{3p,-a' p, = {3' ;
,
(a -a')-({3 -{3')p, +'Y{3'(T
17=
'Y{3'
Under the modified system the proportion of underweight packets is ct>[(l-p,')/(T'] whence the stated result.
60
UL (1972). The expected payment per page is (in £)
ix e- IL + 1 x p, e- IL +ix [1-(1 + p,) e- IL ], whence the answer. If £0 is the new payment to the proof-reader for every page on which he
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
401
detects no errors, then 9 x e - JL + 1 x,.., e- JL +lX[l-(l +,..,) e- JL ] = 1
whence
9 =i(S+,..,-e- JL ).
61 UL (1973).
co
dx f (1x'+l +X)H4
E(x') =(,\ +2)('\ +3)
o
=(,\+2)(,\+3)B(r+2,'\-r+2), if '\+2>r whence the result. Using the particular results for E(x) and E(x 2 ) gives the value of E(T).
62 UL (1973). To evaluate E(x - a) use the transformation y = log(x - a) where -00< y
f
E(x - (3) =_1_
u../2;, 10&(/3-0<)
[(a - (3)+eY ]
e-(Y-JL)2 /2q 2
dy
= -«(3 -a)[1-(8)]+eJL+~'[1-(8-u)], by reduction of the two components of the integrand into normal integrals.
63 UL (1977). The probability density function of X is k e-,\x 2x 3, where the proportionality constant k = 2,\ 2 is obtained from the equation co
f
k e- Ax2 x 3 dx = 1. o
The integral is readily evaluated in terms of the gamma function by using the transformation z = x.Ji... The average consumption of gas per day is co
f
co
2,\ 2 e-,\x x 4 dx = ~ 2
fe-"u~ du,
JA. 0
o
where u = '\x 2 ,
whence the result. The required inequality follows by noting that the annual open fire cost is aN and the annual cost with gas fires is
3 /7T (3 + 300-y . 4 Vi .
64 UL (1977).
co
E(T) = N - N(l - a) f x . e-tx2-/3x dx o
=N -
N(1- a) e!/3 2
f /3
(z - (3) e-!z2 dz,
where z = x + (3.
402
EXERCISES IN PROBABILITY AND STATISTICS
Note that
I e-~z2
I
00
z
00
dz =
e-~(32;
e-!z2 dz =.J2";[I-({3)],
(3
(3
whence E(T). For (33 negligible, (3 1 I I I > dt--+-. 1 (3 ({3)=-+- e-,t
2fu o
2fu
Hence E(T) - N - N(I- a) [ 1-
~N -
~ e!(32 . {3 ~; (1- (3 ~;) ]
N(I-a)[ 1-~e~(32
J,
since 0 ~ {3.J2/7T~ 1 and (3.J2/7T(1- (3.J2/7T)- ~~. The stated inequality now follows directly by using the upper limit of E(T) in the inequality T~E(T).
65
UL (1977). If k is the proportionality factor, then
I 00
E(X') = k
X",+,-1
(1 + {3x",)2n dx.
o
The integral is reduced to a beta function of the second kind by using the substitution z = {3x"'. Hence for r = 0, k = (2n -1)a{3. Logarithmic differentiation of the probability density function gives the equation (a -1)(1 + (3x"')-2na{3x'" =0
for the stationary values of X. For the mode the second logarithmic derivative of the probability density function of X is <0.
66 UL (1975).
P(X~a)= B(m, 1 2)
'" IX
Ill -
1
(I-X)dX,
o
whence the result on integration. P(X ~ (3) is obtained by symmetry. Hence P(X~{3 X~a)
P(X~{31 X~a) = '
P(X~a)
67
P(X"':::{3) ~
P(X~a)'
UL (1973). 1
E(X') =! I x'[1 + Ax(l- x 2 )] dx, -1
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
403
I\'hence the result on termwise integration, x
F(x) =
J![1 + AX(1- X2)] dx, -1
whence the stated result. The median is obtained as a root of the equation F(x) = o.
68 UL (1973).
00
E(X2r )=k
x2r+2 k (2r+3 1-2r) J 1+x4dx=2"B -4-'-4-'
if 1-2r>0,
the integral being evaluated by using the substitution u = (1 + X 4)-1, where 0';;; u:;:;;; 1.
Differentiate logarithmically to obtain the modes and the points of inflexion. The modes are at x = ±1 and the points of inflexion are at x::: ±(2 + "'lJ)i; ±(2 -../¥)i.
69 UL (1972).
00
E(xr) = a(3
J(l~;:~)2 dx o
and the integral is evaluated by using the substitution z = (3x"', where O
70 UL (1971). (ii) Use the substitution u = !x 2 , where 0:;:;;; u :;:;;; A. (iii) Note that for fixed r l(r, 6) ~ 2(r-1)/2[[(r+ 1)/2] as 6 ~ 00. Then II
P(Z:;:;;;6>0)=<1>(6)+
~J exp[-!z2(z4-6z 2+3)]dz
nv27T 1 = <1>(6) + r.::- [1(4, 00) - 61(2,00) + 31(0,00)] nv27T 1 + r.::- [1(4,6)-61(2,6)+31(0,6)], nv27T -00
whence the result on reduction. UL (1970). Since the probability density function of X is a function of Ixl and the distribution is symmetrical, all odd central moments of X and E(X) vanish. If k denotes the proportionality factor, then
71
00
E(~r) =
Jx2r(a + Ixl} dx 2k[a J e- x 2r dx + J e-l3xx2r+1 dx J. k
e- 13txl
-~
=
~
~
13x
o
0
whence the stated result on integration using gamma functions. For r = 0, E(X2r ) = 1 gives k = (32/2(a(3 + 1).
404
EXERCISES IN PROBABILITY AND STATISTICS
72
UL (1969). The probability density function of X is an even function of x and . E(X) = O. Hence So 00
iJ.2r(X) = E(X2r ) = k
J
(a + X 4 )X 2r e- x2 dx
00
J
= 2k e- (a + X 4 )X 2r dx, X2
o
where k is the proportionality factor. The integral is evaluated by the use or the substitution x 2 = z. For r = 0, E(X2r ) = 1 gives k = 4/(4a + 3)~. Logarithmic differentiation shows that for a> 1, the mode of X is at the origin. For a =~, var(X) = 1, 1'2(X) = -~ and maximum ordinate is 3/4/;. For a unit normal variable Z, var(Z) = 1, 1'2(Z) = 0 and maximum ordinate is 1/~ which is less than the maximum ordinate of X. But since 1'2(X) < 1'2(Z) , the frequency curve of X is to be "defined" as platykurtic. 73
UL (1968). The probability density function of X is e- x/(1 +e- X )2.
1
=
Jyt(l- y)-t dy o
=B(I+t, I-f),
Itl<1.
E(e tX )
The moments are obtained by expansion of as a power series in /, whence 1'1 and 1'2' (i) P(X:::; x) = ~ and P(X:::; x) = i give the first and third quartiles as -log., 3 and log., 3. Hence the S.I.Q.R. (ii) Logarithmic differentiation leads to the equation e- 2x -4e-x +l=0 for the points of inflexion. 74
UL (1967). If k is the proportionality factor, then
since the distribution is symmetrical about the origin. Hence 00
00
k[J~J~+ l+x2 l+x4
J dX]=1 00
2
x l+x4
.
0 0 0
The second and third integrals cancel out as can be seen by putting x == 1/11 in the third integral. Hence k = 2/7r.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
405
Logarithmic differentiation leads to the stated equation for the turning The equation can be written as (v2+~v+i)(v-~)=k where v=x 2.
~alues.
rhus v -~. The modal values are approximately x = ±~, since the second derivative is <0. The improved approximation is obtained by setting v = 8 + ~ where 8 is small. This gives (8 +~)3_(8 +~)-2 = 0, whence neglecting 1;2 and 83 we have 8 =;k,. Hence the improved approximation for the modal values.
7S UL (1967). 00
J[exp[-2ax/f3+tx]+ (1-a)2 exp[-2(1-a)x/f3+tx] ] dx
2a 2
E(e IX )=13
~
o 00
= 2a 2 J[exP[-Z(2a - f3t)] + o
e
:ayeXP[-z{2(1-a)-f3t}]] dz
where z = x/f3 and 2a - f3t > 0, 2(1- a) - f3t > 0. The stated result follows on integration. To obtain the moments, note that log Mo(t) = log[ 1- 1 ;~~(~:~) f3t]-log[ 1- :~]-log[ 1- 2(:~a)] =f3t+f3
2[1-2a(1-a)] t 2 2a(1-a) 2+""
on expansion,
whence the E(X) and var(X). The lower limit for var(X) is obtained by noting that a(1-a)~~. 76 UL (1966). The moment-generating function of X is E(etX ) = ell" whence the moments. The proportionality constant k for the probability density function of Z is obtained from the equation 00
k
Je-!z,[1+az(z+1)]dz=1,
whence
k
=
1 .Jh(1+a) .
The vth moment of Z about the origin is
J 00
E(ZV)
1 fu(1+a)
= IL~(Z).
e -iZ2 (z 1'+ az 1'+ 1 + az 1'+2) dz.
-00
406
EXERCISES IN PROBABILITY AND STATISTICS
For v = 2r + 1 and v = 2r, the required moments are obtained by using the integral expressions for the moments of the unit normal variable X. In particular, E(Z) = a/(1 +a) and var(Z) = 2-1/(1 +af. 77
UL (1965). 00
J
E(e'X) = A e-(A-')x dx = (1- t/A)-I. o
The cumulants of X are obtained by the expansion of -log(1- t/A). The point-probabilities of Yare c e- Ay where 00
c
L e-A'=1
or c=1-e-A.
• =0
Hence 00
G(6)=(1-e- A)
L (6e-A)' =(1-e-A)/(1-6 e-A) . • =0
Therefore the cumulant-generating function of Y is log(1-e- A) -log(1-e- A e') whence the first two cumulants on expansion. Finally, e A -1-A Kl(X)-Kl(Y) = A(eA -1); K2(X) - KiY) =
(e A - 1f - A2 e A A2(eA _1)2
The stated approximations are now obtained by expansion. 78
UL (1972). P(X = r) = Probability that there is one ace in the first r cards x (r+ l)th card dealt is second ace
The recurrence relation is obtained by considering the ratio P(X = r + 1)/P(X = r). It then follows that P(X = r + 1) > P(X = r) if (49-r)(r+1»r(51-r) or if r<4j. 79 UL (1977). For r>O,
whence the mean and variance of X. The likelihood of the n observations is
" X~ ( 1)" exp (1" -i i~ ) Il
L = 6A 4
Xi
•
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT SO
407
that nx
L 3 log Xj, II
log L = constant - 4n log A- - +
A
j=t
whence the estimate A on differentiation with respect to A. We have E(x) = 4A; var(x) = 4A 2/n = E(x 2) - E2(X). Therefore E(x 2) = 16A 2(1 + 1/4n). Hence an unbiased estimate of A2 is nx 2/4(4n + 1). Also, E(x 2/16) = A2(1 + 1/4n).
80 UL (1973). The likelihood of the sample observations is
( 1)11 .exp--1 L (xj-Of II
L= - SO
that
~2wO
20
j
=t
n 1 log L = constant-llog 0 - 20 [ns 2 + n(x - Of].
Differentiation with respect to 0 leads to the stated equation for 0. The quadratic in 0 has one positive r
81 UL (1972). Use the transformation z = x/O to evaluate E(xr) by a straight application of the gamma function. The logarithm of the sample likelihood is logL=constant-3nlogO+2 whence the stated value of
0.
II
nx
j=t
0
L logxj--
Var(O) = var(X)/9n.
E(L t ) = ~ 4n
r
E(xD = !E(X2)
j=t
and E(L2) = 1E(x2) = 1[var(x) + E 2 (x)], whence the stated results. 82 UL (1971). (i) P(X = n) = Probability that there are exactly r - 1 white balls in the first n - 1 balls sampled x Probability that the nth ball is white =
(ii)
E(OX) =
(nr-1 -l)pr-l(l_ p )II-r X p. i: Oll(nr-1 -l)prqll_r
lI=r
=
=
r
(E.)r (n ~ l)(qO)1I q lI=r r 1
~ (r+s-1) (pO)' s~o s (qO)S = (pO)'(l- qo)-r.
408
EXERCISES IN PROBABILITY AND STATISTICS
(iii) The log likelihood of the sample is log L
=
constant + r log p + (n - r) 10g(1- p)
whence p and var(p) are obtained by a straight application of the method of maximum likelihood.
I
(iv) E(r-1)=pr (n-2)ql1_r n=r r-2 n -1
(r+S-2) qS = pr(l_q)-r+l = p.
~ = pr l.J
S
s=O
83 UL (1969). The point-probabilities of the truncated Poisson distribution are
so that E(nr )
=
NP(X = r)
and
1 <X> E(IA- *) = - L rE(nr ) N r=2
=
IA-, on summation.
A simple argument for the derivation of var(1A- *) is as follows. Let Y 1 , Y 2 , ••• , YN be random variables denoting respectively the number of prosecutions each motorist has. Then Yj are independent random variables, each having the same distribution as that of X. Next, define new random variables Zj such that Zj = Yj if Yj;;.2 and Zj = 0 otherwise. Then N 1 NIA-* = j~ Zj and var(1A-*) = N var(Zj), where
1 <X> rlA- r E(Zj) =-IL- L -, = IAe -1 r=2 r. and
Hence var(1A- *) as stated and var(,l) = var(1A- *)
(elLelL-1)[1_ (_1A-_)2] elL - 1 ~ 1
as IA- ~ 00.
Note that IA- * is more easily calculated than ,1 and for large IA-, IA- * may be preferred to ,1 as a quick estimate of IA-.
84 UL (1966). The log likelihood of the sample observations is 11
log L
=
-n log '7T -
L 10g[1 + (Xj -lA-f], j=1
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
409
whence the equation for (l. is
n 2' the integral being readily evaluated by the use of the transformation x -IL = tan 0, -7T/2 ~ 0 ~ 7T/2.
8S UL (1977). The log likelihood of the sample observations is
n n 1 n log L = --log(27Tk) --log 0 - (~- 0)2 2 2 2kO i =1
L
whence the equation for 6 on differentiation. Since of the quadratic equation gives the value of 6. For the variance, observe that n 2 -2 d 2 logL n d02 202- k0 3 (s +x)
n
=20 2-
6> 0,
the positive root
1" k0 3 i~ x'f.
For the limiting variance put k = ILIO, whence 21L2 1 IL var(O) = - . 2 k 2 -+ - as n IL+ n A
k -+ o.
86 UL (1973). The least-squares estimate of IL is obtained by minimizing
n = L"
(n - v + 1)(x -IL f
with respect to IL,
v=l
whence IL*=
2 " 1) (n-v+1)x". n n + v=I
L
(
Clearly,
4
"
var(1L *) = 2( 1)2 L (n - v + 1)2 var(Xv) n n+ v=1 and 36 " 1)2(2 1f (n - v + 1)4 var(Xv), var(T) = 2( n n+ n+ v=l
L
410
EXERCISES IN PROBABILITY AND STATISTICS
whence
2
2a var(IL*)= n(n+l);
87
9a 2 var(T) = (211 + If .
UL (1972).
1 E(a*) = -
n
E«(3*) =
/I
n v=!
v=l
X)2;
t
1 . (Xv - x)E(yJ (Xv-xf v=!
I v=l
=
l'
/I
L E(yJ = a +- L (Xv -
(3 + l'
JI
(Xv - X)3 /
JI
(Xv - X)2
If Xv are equispaced, then I~=! (Xv-X)3=O. Hence
E«(3*) = (3;
E(a*) f a.
If we set X k == I~=l (Xv - X)k for k = 2 and 3, then
E(a*) = a +I'X2 /n;
E«(3*) = (3 +I'X3 /X2
whence
and
88 UL (1971). Observe that rv
=
Yv -a*- (3*(xv -x)
_ [ 1 (xv - X)2]
- Yv 1---
X
n
-
[1
" (xv - x)(Xj - X)] ~ Yj -+ . j"ov n X
Hence E(rJ = E(yJ - E(a*) - (xv - x)E«(3*) = 0 and
- xf]2 2 () a 2[1 - 1n (xv X- X)2]2 + j"ov,,[1-n + (Xv - x)(Xj a X
var rv
~
=
=
a
2[(n-l)2 n -1 (xv - X)4 - - +--+-'---'------,----'---n
+ (x v;2xf +
ttl
2(x -x) {
~X
n
2(n -1)(Xv -
nX
X2
(Xj -xf-(xv -xf}
j~! (Xj-x)-(Xv-x) /I
=a2 [n:l_ (xv;xf].
}]
xf
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
411
89 UL (1970). From general theory 0* = a* - x(3 *, whence E(e*) = a - x(3
and
-2)
1 var(e*) = u 2 (-+~ , 11
where
X
n
X =
I
(Xi _x)2.
i~l
Hence cov(O*, (3*) = E[(O* - 0)«(3* - (3)] = E[(a* - a)«(3* - (3)]- xE«(3* - (3)2
XU 2 X· When the Xi are the
11
natural numbers, then n
I
x=!(11+1);
X~ = kl1(11 + 1)(211 + 1),
i=l
whence
* * - [ 3(n + 1)
J!
corr (e ,(3 ) - - 2(2n + 1) . 90
UL (1967). E(T- /.Lf = C 2 x E[Jl (Yi - /.L)2+ 2(I1C; 1)/.L Jl (Yi -IL)+ (nC~!)21L 2J =
nC 2u 2+(nC-1)21L2.
Hence the stationary value is C = l/(n + v 2 ).
91 UL (1965). Clearly, E(x) = E(y) = IL; u2 var(x) =-;.
n
var(y) =
u2
-
n
[1 + (n -l)p].
Now T is an unbiased estimate of IL if a + (3
=
1. Also
u2 var(T) = - [(1- (3)2+ (32{1 +(n -l)p}].
n
Hence for minimum var(T) u=
1+(n-1)p . 2+(n-1)p'
(3 =
1
-,------
2+(n -l)p .
Since the Yi are equi-correlated, -[l/(n -1)] os:; p os:; 1, so that (n-1fp2 4{1+(n-1)p}
-,,-'----,--'--'--,- ~ 0
.
Hence for Pt=O. varG(x+y)]>minvar(T). Note that x and y have unequal variances and so the best estimate of IL is T.
412 92
EXERCISES IN PROBABILITY AND STATISTICS
UL (1968).
If x is the unknown (n + 1)th independent observation of X, then x - i . normally distributed with E(x - x) = 0 and var(x - x) = (n + 1)0-2In. Henc~s
irrespective of the magnitude of fJ.
'
x-x
has Student's distribution with n -1 d.f. whence the stated confidence interval for x.
96 UL (1969). Under normality assumptions of the y observations, it is clear that Yk is normally distributed with mean a + /3(xo - x) and var(Yk) = 0-2/k. Therefore Yk - & -/3(xo - x) is normally distributed with mean zero and variance 1 1 (Xo-X)2] . 0- 2 [ k+-;;+ X . Hence the ratIo A
Yk -& - t3(xo-x) 1 1 (xo- X)2]i s [ -+-+..:......:'----~ k n X has Student's distribution with n - 2 d.f. This now leads to the stated confidence interval for Yk.
94 UL (1971). Here x - L is normally distributed with mean zero and variance 0-2[(1In)+1-2A(1-A)]. Hence the ratio
x-L
S[~.+ 1-2A(1- A)] has Student's distribution with n -1 d.f. This leads to the stated confidence interval for L. The lower limit is obtained by noting that 0,..;;; A(1- A) ~!.
9S UL (1975). Observe that x - Xo is normally distributed with mean zero and variance 20-2/n. Hence the ratio x-Xo
s../2frt has Student's distribution with n -1 d.f. This leads to the stated confidence interval.
96 UL (1966). The rth factorial moment of w is E[w(r)]= n(r)pr, for
r~O.
Therefore fJ.3( w) = E[( w - np )3], whence the result on taking expectations after expansion. Since fJ.3(w)=n(p-3p2+2 p3), an unbiased estimate of this is _ [W 3W(2) W(3)] T- --~+2---cJ) , n
n
n
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
413
which gives the answer on reduction. Observe that E(p*) = p; var(p*) = p(1- p )/n, and for large samples dT)2 var(T) = ( d * p
x var(p*)
p*=p
whence the stated result. 1)7
UL (1968). The log-likelihood of the sample observations is
log L = constant+ nl log(2 + 6) + n210g(2- 6) + n3 log(1- 6) + n410g(1 + 36) whence, on differentiation, the equation for 6 is n3 + 3n4 =0. 1-6 1+36 Also, var(6) is readily obtained from d 2 10g L/d6 2. With misclassification, the log-likelihood is
nl 2+6 A
n2
_
A
_
A
A
2-6
log L = constant + nllog(2+ 6) + (n2 + n3) log(3 -26) + n410g(1 + 36) whence the equation for 6* is nl 2(n2+ n 3) + 3n4 =0 2+6* 3-26* 1+36* . For large samples var( 6*) > var( 6) if 4(3 - 26)(5 + 26 - 4( 2) > (2 - 36+ ( 2)(29 + 36) which reduces to (6+2)(36+1»0.
98 UL (1969). The log-likelihood of the observed sample is log L = constant+ al log(2- 6) + a210g(1 + 6) + a310g 6 + a410g(1- 8) whence, on differentiation, the equation for 6 is _~+~+ a3_~_0
2-6 1+6
E(X) = N(1-26)
8
1-8- .
and var(6*) =
1 + 26(1- 6) 2N '
obtained by using the properties of the multinomial distribution. The limits for var(8) follow since 0~8(1-8)~~.
99 UL (1970). Since I~=l mi = N, a constant, therefore n=l dmJd6 = O. The probability of obtaining the observed sample is L
=
kN!
IT
i=l
.
Ii (mi)
a,
.1 i=l N a,.
and the equation for 6 is obtained by logarithmic differentiation.
414
EXERCISES IN PROBABILITY AND STATISTICS
Note that E(X) = 0, var(ai) = mi(1- m;/N) and cov(ai> a j ) == -m.rn/N
i =1= j. Hence var(X) is obtained by using the fact that X is a linear functio~ of
the ai.
100 UL (1966). (i) Clearly, var(x) = :2
[tl var(Xj) + 2 ~t: cov(Xj, Xj+l) J,
whence the stated result. (ii) E(S2) =
~ E[ f. 1
n
i=1
1 [ =---:-E n
=
n
1
(Xj -
X)2]
L x7-nx 2 11
]
i=1
~ 1 [.f. E(x7) 1=1
nE(x 2)]
[.f.
~ 1 var(Xj) - n var(x)], since E(Xi) = E(x) 1=1 n whence the result. =
(iii) This follows by noting that var(x);;:. 0 for all nand E(S2);;:. 0 for all n ;;:. 2. Note that var(xw )
=
[f.
2( 4 )2 i 2 var(Xj) + 2 L ij cov(Xj, ~)] n n+1 i=1 i
=
2(
n n+1
[
)2
11
11-1
L f+2p i=1 L i(i+1) i=1
]
,
whence the result on summation. Eff(iw) = var(x)/var(xw ) 3(n+1) 1+2p(n-1)/n 3 ~ as n~oo. 4n ·1+2p+(1-4p)/2n 4 101 UL (1967). (i) Since var(X - Y) = var(X)+var(Y)-2 cov(X, Y);;:'O, therefore cov(X, Y) .::;;~[ var(X) +var(Y)], whence the stated result. (ii) Suppose that E(X) = 6 1 and E(Y) = 62. By definition, cov(f , X) = E(Y) - E(f)E(X) ( Y - ( 2) + 62 ] 62- 6 1 • E [ (X - ( 1)+ 61 X - 6 1 Y - 62 (X - ( 1)(Y - ( 2 )] - 62-62 . E [1 --6-1-+~6 162 ' whence the result on taking expectations. =
102 UL (1968). Assume that for two random variables X and Y, var(X) = O"f;
var(Y) = O"~;
cov(X, Y) =
pO"10"2.
415
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
Then var(X ± Y) = (0"1-0"2f+2(1 ±p)0"10"2;;:'0 for all 0"1 and 0"2' Hence for 0',==0"2,1±p;;:'0.
1 [" (i) var(i) = n2 i~1 var(Xi)+ i~ COV(Xi, Xj) ] , whence the result on substitution. (ii) Here Yi = Xi - i = (1-1.)Xi -1.
n
L Xj'
n j""i
Therefore var(Yi)=
(1-~r var(x;)+(~)\n-1)Var(Xj)
whence the result on substitution. (iii) Put Zi = Xi - /.L and
z =1. i n
i~1
Zi' Then
Expand and take expectation over the z's whence the answer.
103 UL (1969). Here S" = Ill=2 Zi where P(Z, = 1) =i;
P(Z, =0) =~ for 2~r~ n.
Therefore E(Zr) = t cov(Z" Z,+l) =
--16;
var(Z,) = /i;; cov(Z;, Zj) = 0
for
Ii - il ;;:,2.
Hence E(S,,) and var(S,,).
104 UL (1970). (i) If E(X) = 81 and E( Y) = 82, then cov(X + Y, X - Y) =E[{(X - 8,)+(Y -( 2 )}{(X -( 1)-(Y -82 = E[(X - 8 1f- (Y - ( 2)2] = var(X) - yare Y), whence the result. (ii) By definition,
m
whence the result on summation of the double series.
416
EXERCISES IN PROBABILITY AND STATISTICS
lOS UL (1971). (i) Let E(X;) = 8i, i = 1, 2, 3. Then, by definition, cov(X1+ X 2 , X 2 + X 3 ) = E[{(X1- 81) + (X2 - 82 )}{(X2 - 82 ) + (X3 - 83 )}] = E[(X2 - 82)2], since the X; are uncorrelated =var(X2 ). (ii) Here E(Si) = 0, i = 1, 2, 3. Hence cov(Sl> S2) = =
r
"
r
"
L L E(x.Xt) .=11=r+1 L L
.=11=r+1
L L r
=u 2
cov(x., Xt) "
pl-.,
.=11=r+1 whence the result on summation. Cov(Sl> S3) is obtained in a similar manner.
106 UL (1973). Since E(Yi) = 0, we have 1 i+,,-1 i+r+n-1 cov(Yi> Yi+r) =2 L L E(x.Xt).
n
s=i
t=i+r
If r;;:;' n, then E(x.Xt) = 0 for all permissible values of sand t (s 1= t). Therefore cov(Yi> Yi+r) = O. If r:OO;;;n-1, let n-r-1=k, where k;;:;.O. Then 1 j=n-1 cov(Yi> Yi+r) = 2 E(xr+j), the cross-product terms being zero n j=n-k-1
L
107 UL (1966). For the Poisson distribution P(X = x) = e-"',A,x/X! whence QO
E(e'X ) = e-'"
L (IL e1nx! = exp lL(e' -1). x=o
It then follows that
Hence E(e 'Z ) = e-',/i;; . E(e'x Jn7j;.) = e-',/i;; . exp[nlL(e'/,/i;; -1)]. Logarithmic expansion gives the result for K(t).
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
108
417
UL (1967). 00
L (A8Y/r! = e
E(8 X ) = e- A
A (9-t).
r=O
Therefore G(8l> ( 2) == E(8J' . 8n = E[(8l82)x']E(8~2) = exp[A l (8 l 82 -1) + A2(82 -1)].
The probability-generating function of Y is G(1, ( 2 ) whence P(Y = j). The probability-generating function of the conditional distribution of Xl given Y= j is G(8 l I Y = j) = coefficient of 8~ in G(8 h = (1 + p8 l )i/(1 + p )i.
(
2)/P(Y = j).
On expansion the coefficient of 81 gives P(XI = r I Y = j). 109 UL (1967). E(8 X )
" (n). (p8)iq"-i = (q + p8)".
=L
i=O
]
Hence G (81) ( 2) = E[(0 1 (2)X,] . E[ 0~2] =
(ql + PI 81( 2)"'(q2 + P2 ( 2)"2.
The probability-generating function of Y is G(l, ( 2 ) whence P(Y = r), and the probability-generating function of Xl given Y = r is G(OI I Y = r) =
[arG(O~, ( 2)] a0 2
jP(Y = r), 9 2 =0
whence, on expansion, P(X1 = slY = r). 110 UL (1967). The probability-generating function of X is 00
G(O) =
L Pror, r=O
and the factorial moment-generating function of X about the origin is Ho(a) == G(1 + a) r=O
f t r(s) a;s. = f a; f s.
=
Pr
r=O
s=O
s=O
r=s
Prr(s)
418
EXERCISES IN PROBABILITY AND STATISTICS
Again, putting () = 1 + a, we have co
co
«()-1)S
r=O
0=0
s.
L Pr()r = L - ,=
f.L(s)
t
f
f.L~0)
co
S. ()k
(S)()k(_l)S-k k=O k co (-1y-k
co
()k
co
0=0 =
k~O k! S~k f.L(s) (s-k)!
=
L -k' L -.,- f.L(k+j) k=O 'j=O J.
(-1)j
Equate coefficients of ()' on both sides to get Pro The particular result follows on summation of the series.
111 UL (1966). Here E(ri) = P1;P; var(ri) = P1;pq; cov(rj, rj ) = 0 if j. It then follows that E(x)=E(y)=p and var(x) and var(y) are also readily obtained. Note that E(e'X ) = (q + P el/N)N whence E(e 'W ) = e-P1/
n (q + P e l/",k)", i=1 k
E(e 'Y ) = and
E(e 'Z ) = e-P1/
n k
(1 + P el/n,kuo)".
i=1
Logarithmic expansion of E(e"Y) and E(e 'Z ) now leads to asymptotic normality. The approximation for var(x)/var(y) is obtained by noting that
t ! t ~ [1 _(ni ~ ii) + (ni ~ ii)2 _... ] -~ [1 + V~~i)J. =
i=1 P1;
i=1 n
n
n
112 UL (1966). The sample of n unrestricted values can be selected in (~) ways. Next, if
x and yare the largest and smallest integers in the sample, then the choice of the remaining n - 2 is restricted to x - y -1 numbers in all. Hence
where, evidently, n~X~N and for given X, ¥;;;;.1 and X-¥-1;;;;'n-2, whence the stated conditional limits for ¥. Clearly, for fixed X, n -1 ~ Z ~ X -1 so that the joint distribution of X and Z is
P(X=x,Z=z)=
(:=~)/(~
where
n~X~N.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
419
Hence the marginal distribution of X is
P(X=x)=
L
x-I
(
z=,.-1
z- 1)j(~ n-2 n
where
xf'
Xf z=,.-1
(Z-1)= (U+n-2) n -2 ,,=0 U x-n
=
L
coefficient of 0" in (1 + 0),,+,.-2
,,=0
=
{1 O}" 0 {1 + O}" in (1 + 0),.-2 L - -
x-,. L coefficient of 00
,,=0 =
in (1 + 0),,-2 -+X-,.
coefficient of 00
,,=0
0
= coefficient of 00 in (1 + O)X-l • O-(x-,.l
=\x-n (X-1)= (X-1). n-1 Again, for fixed Z = z, we have z + 1 :s;;; X :s;;; N, whence the marginal distribution of Z is P(Z = z) =
f (Zn =21)j(N\n)
x=z+1
=(N-Z)(:=~)j(~, with the stated limits for Z.
113 UL (1966). The probability that a prosecuted motorist has x prosecutions is e-.... '.c.x/(1-e-.... ). x!,
1:s;;;x
It now follows that the rth factorial moment of x is 00
e
1-e ....
(rl
r
L~
-....
E[x(rl ] =
x=1
x!
f.Lr
-1-e-..... Hence E(x) = _1£_ .
1-e-.... '
Therefore var(x) -1 1£ e-.... E(x) - -1-e-.... ' which is P(x ;;;:.2) for the truncated distribution.
420
EXERCISES IN PROBABILITY AND STATISTICS
114 UL (1966). Consider first the joint distribution of Xl> x,,+l and X Zv + 1 and then the conditional probability of the 2r observations. It then follows that p
=
(2v+1)! . (2r) {(v-1)!F r
Xv + 1
J J J[ Jf(x)dx ]v+r-t 00
-00
00
Xl
00
Xl
X v +l
X,,+l
To evaluate the triple integral make the transformation Xl
U
=
X"+l
Jf(x) dx;
J
v=
X2v+l
f(x)dx;
W=
f
f(x) dx
where O
115 UL (1966). By definition, «I>(x) =
k 1 2
X
f e-!u 1
f
J2;
0
2
du
_1
2
=-+-- L.e "u du z)' f L'" (-!u --du x
1 1 =-+2 J2;
r!
r=O
o
whence the result on integration after changing the order of summation and integration
f J2; '"
E[etIX1 ]=_1_
=
(2)! f'" 1T
etlxl-!x2
t
1
eX-"X
2
dx
dx
o =
(!)! j
e!t 2 e -!<X-t)2 dx o
=
2 e!t 2 «1>(t)
Logarithmic expansion gives the cumulants.
116 UL (1966). Set
Vt
= Xl/a and
Vz = xz/a.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
421
The probability density function of the joint distribution of VI and V2 is n 2(n -1f(VIV2)n-2(1_v l )(1- V2), O::s;;; VI> V2::S;;; l. To evaluate the distribution of u, put VI = UV2, V2 = V2 and consider the cases O
117 UL (1966). If k is the proportionality factor, we have
ff I
E(xrys) = k
00
e- y (l-x)x ry",+s(1- x)"'+~ dx dy
o
0
I
=
00
k f xr(1- X)~-'-l dx f e- y (l-x){(1- x)y}",+s(1- x) dy o
=
0
kB(r+ 1, (3 -s)f(a +s+ 1).
For r = s = 0, E(xrys) = 1 whence k = (3tr(a + 1). The probability density function of the distribution of X is 00
{3
f{a+1)
f
(1- X)~-l e - Y(I-X){y(1- x)}"'(1- x) dy = (3(1- X)~-l
,
for O::s;;;X::s;;;1.
o
The probability density function of the distribution of Y is
f(:+
1
1) . e-Yy'" f e YX (1o
x)"'+~ dx.
This may be put in the required form by use of the substitution y(1- x) = t.
118 UL (1967). Clearly,
Hence by substitution in the probability density function of X, the stated distribution of Y is obtained. E(Y) = 24 a4
f'"x 2(a-x)(a 2-x 2)idx. o
The integral is evaluated simply by using the substitution x = a sin O.
'"
P(Y::S;;;a) =~ f[2a(4a2_y2)i-(4a2_y2)]y dy. 4a o
Evaluate the integral by using the substitution y = 2a sin O.
422
EXERCISES IN PROBABILITY AND STATISTICS
119 UL (1967). co
E(e iIX ) =
J
J:...
eilX-lxl/a
20'
[J
dx
J
00
=
~
00
e-(1+ila)"
o
E(e i1z )
du +
e-(1-ila)"
du ]
0
1 1 + t 20'2' = E[eil(1-.... )y']E[eit....y2] 1 1 1 + t 2(1_1J.)2 . 1 + t 21J. 2 1 [(1-1J.)2 =1- 2 1J. 1+t 2(1-1J.)2
IJ. 2
]
1+t21J.2·
The sampling distribution of z now follows directly by the use of the Inversion Theorem.
120 UL (1967). (i) From the definition we have cOV(Xj_k, Xj+2) = a cov(Xj+l> Xj-k) + (3 cov(Xj, Xj-k) so that Pk+2 = apk+I + (3Pk' (ij) Similarly, var(Xj+2) = a 2 var(Xj+l) + (32 var(Xj) + 2a(3 cov(Xj, Xj+I) +var(ej+2) so that var(e) = 0'2(1- a 2- (32)-2a(30'2pl> where from (i) for k = -1, PI = apo + (3P-l> Po = 1, PI = P-I' Hence PI = a/(1- (3) and var(e) follows. (iii) The general solution of the difference equation in (i) is Pk = AA~ + BAt
with Po= 1,
PI = a/(1-a)
Use of the initial conditions now gives
A
=
-A2+ a /(1-(3). Al - A2 '
whence Pk'
121 UL (1967). Note that 1 1 z v=-tan- 21T W It then follows that a(u, v)
--=
a(w, z)
1 [I( 2 2)] --exp -2" W +z 21T
423
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
whence Wand Z are independent N(O, 1) variables. Therefore W+Z2= - 2 loge U is distributed as x2 with 2 d.f. If Ul> U2, ... ,Uk+2 are independent realizations of U, then n=1-2Ioge Uj is distributed as x 2 with 2k d.f. Also (-2 loge Uk+l)~cos(21TUk+2) is an independent N(O,1) variable. Hence w2+ L~= I - 2 loge u; is distributed as X2 with 2k + 1 d.f. 122 UL (1967). The probability density function of the distribution of XI and x,. is
4n(n-l) 2 211-2 ,,::: ""'">211 (X I1 -Xl) XIX,,, O-"Xl-.,a;x,,"'-XI· a Hence the probability density function of the distribution of
n(n-1)
----'-.........,,--'- (UV)"-2(U 2 - v 2) 2a 2n "
0 ~ V ~ a' v ~ U ~ 2a - v
For the marginal distribution of v, integrate over v~u<2a
U
and v is
U
•
with the limits
-v.
The marginal distribution of u is obtained in two parts. For 0 ~ u ~ a, integrate over v in the range 0 ~ v ~ u, and for a ~ u ~ 2a integrate over v in the range 0 ~ v ~ 2a - u. 123
UL (1967). y
co
P(X~Y)= J Jf(x,Y)dXdY=~ al+~l o
0
so that the normalized joint distribution of X and Y subject to the condition that X ~ Y has the probability density function ~2(al
+ ~l) exp[ -~2Y -(al + ~I -
~2)X],
O~X ~ Y; O~ Y <00.
(i) The marginal distribution of X has the probability density function co
~ial + ~l) exp[ -(al + ~I - ~2)X] J e-/32
Y
dy = (al + ~l) e-
x
O~X
Oi) The required density function is ~2(al + ~l) exp[ -~2Y - (al + ~l - ~2)x]/(al + ~l) e-
The conditional probability is P(x
~ Y ~ 2x) =
2x
f ~2
e-/3,tY-x) dy = 1-e-132x.
x
UL (1968). Consider first the joint distribution of Xl> Xv+l and X2v+l and then .t~e conditional probability of the 4r observations. Then the required probabIlity
124
424
EXERCISES IN PROBABILITY AND STATISTICS
is (2v+1)! (4r)! p = {(v-1)!Y(r!)4
Joo Joo Joo [JXI
f(x) dx
-00
XI
X ... +l
X 2 ... + 1
X [
-00
]r[ XJ"+I
]r+V-l f(x) dx
XI
00
J f(x) dxr+V-l[ Jf(x) dx Jf(X1) dxd(x +l) dX +d(X v
v
2v
+1) dx 2v +1•
To evaluate the triple integral make the transformation
XI
u=
Jf(x)dx;
X 2 ... + 1
X ... +l
v=
J f(x)dx;
Jf(x)dx,
w=
O
Integrate successively over wand v by using the substitutions 1- w = (1- v)t and 1- v = (1- u)z. The integrand in u is easily evaluated as a beta function, whence the stated result.
125 UL (1968). The probability density function of the joint distribution of X and nY2/2 is _1_ e-x2/2a2x _1_ e-ny2/2(ny2/2)n/2-I, -oo<x <00;
ufu
0:0;:;
y <00
r(n/2)
Make the transformation x/u=rcosO; y.vn=rsinO, where O:O;:;r
1 (. 0),,-1 B(nI2, !)' sm . Clearly z n d.£.
=.vn cot 0 whence it follows that Z has Students' distribution with 00
EX,Y[cP(X+CY)]=.l;
00
x+cy
J J[ J -00
0
e- U2/2 dU]t(x,y)dXdY
-00
where f(x, y) is the joint density function of X and Y =Ex.y[pu{U:O;:;x+cy I X=x, Y= y}] = px.y,u(U:o;:;X+cY)
=Px.y.u[y~:O;:;~J. But since (U - X)/(l + ( 2 )! is N(O, 1) and independently distributed of Y, it follows that Z and (U-X)/(1+u 2 )! have the same distribution.
126 UL (1968). The cumulant-generating function of X is 00
K(t) = f.L(e' -1) = f.L
L trlr! r=1
425
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
so that
Kr
= /.t for
r;;'
1. Hence /.tiX) =
K4
+ 3K~ = /.t + 3/.t 2 •
Next,
n
E(e'X) = " E[e'x/,,] i=1
= [exp l1-(e'/" -1)]" = exp n/.t(e'/" -1). Therefore the cumulant-generating function of x is n/.t(e'/" -1), whence E(x) = /.t,
_
var(x)=~,
/.t
3/.t 2
n
n
/.tiX)="3+-2 .
n
E(T) can be readily verified and to obtain var(T), it can first be shown that n-3)]2 911-2 18/.t [T- E(T)]2 = 9(x -11-)4+ [ 6/.t + ( -n(x - /.t)2+ 7 -----;- (x - /.t)2
+ 6 [ 611- +
(n : 3) ] (x - /.t? - 6; [ 6 /.t + (n : 3) ] (x - 11-),
whence the result on taking expectations.
127 UL (1968). Observe that the region of integration In the (u, v) plane has the boundaries u ± v = 0 and U ± v = 2a. Hence (i) for fixed v in 0 < v < a, v ~ U ~ 2a - v ; (ii) for fixed v in -a < v < 0, -v ~ U ~ 2a + v. It then follows on suitable integration that the probability density function of v is a-v -2- for a
O~v
a+v -2- for a
-a~v::;;;O.
P(v;;./3) is obtained directly by integration and so also E(lv \) and E(lvI 2) = E(v 2 ).
128 UL (1968). (i) Clearly, Ur-l - Ur is the probability that X is realized at the nth trial. Therefore the required expected number of trials is 00
L r(Ur-1 r=1
00
ur )
=L
,=0
Ut·
(ii) Here Y,,-1 - y" = probability that the pattern is realized for the first time at the nth trial
= y,,_ 4 X qp 3 whence the difference equation for y", n;;' 4, with the initial conditions YO=Yl=Y2=Y3=1. The generating function for this difference equation is
and y" is the coefficient of zit in the expansion of
426
EXERCISES IN PROBABILITY AND STATISTICS
The expected number of trials required for the realization of X is 00
L
Yr = ep(1) = l/qp 3.
r=O
129 UL (1968). Here Xl> X2, sf and s~ are all independently distributed and F=sVs~ has the F distribution with n -1, n -1 d.f. Now Xl + FX 2 w= 1+F
and
E(wIF)=p.,
whence E( w) = p.. Also, 1 var(w IF) = (1 + F)2 var[(xt + FX2) I F]
u 2 1+F2 =-;;-. (1 + F)2· Hence
where
J 00
{ F} EF (1+F)2 =
1 p"-1)/2 dF (n-1 n-1) (1+F)"+1 B -2-'-2- 0
n-1 4n . 130 UL (1968). The distribution of W is obtained directly by using the transformation suggested. It then follows that w = t/.Jii, where t has Student's distribution with n d.f. Also, W
=!(XI/X2- X2/XI) =~
(.JF - Jp).
where F has the F distribution with n, n d.f. Hence
whence
np2-(4t2+2n)F+n =0. The positive root of this quadratic gives the required answer since for F> 1, t>O.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
427
131 UL (1968).
f 00
(i) E(e itX ) =!
eitx-Ixl dx
00
=
fe
-x
cos tx dx
o = 1- t 2 E(e itX ).
(ii) U j and U 2 are independent N(O,!) variables. (iii) We can write
W = a[(Zl + Z2)2 - (Zl - Z2f + (Z3 + z4f - (Z3 - Z4f] where !(Zl ± Z2) and !(Z3 ± Z4) are independent N(0,1) variables. Hence W = !(xi - x~) where xi and x~ are independent X2 variables each with 2 d.f. Finally,
f fexp[itw -!xi-!x~] dxi dx~
E(e itW ) =!
o =
0
ff
exp[-(1- it)vj - (1- it)V2] dVl dV2
o
0
132 UL (1968). P(X = r) = Probability that the first r cards include any two aces and r - 2 non-aces x Probability that the (r + 1)th card is an ace
(~)(,~82)
2
(5 2) r
. 52-r'
for 2,,;;; X,,;;; 50.
P(¥ = s ! X = r) = Probability that s non-aces turn up after the third ace x Probability that the (r + s + 2)th card is the fourth ace
Hence
P(X=r, ¥=s)=P(X=r)P(¥=s!X=r). 12r(r-1) 52(4)
EXERCISES IN PROBABILITY AND STATISTICS
428 Therefore
50-.
P(¥=S)=
r~
12r(r-l) 52(4)
--'-------;-:-;--- • f or 0 :0;;;; ¥
4(51-s)(3) 52(4)
:0;;;;
48 •
133 UL (1968). I
E(Z) =
00
Jxf(x) dx + Jtf(x) dx = PILI +qt, I I
E(Z2)=
00
Jx 2f(x)dx+ Jt2f(X)dX=P(0'~+IL~)+qt2, I
whence var(Z).
134 UL (1968). Here var(i) = \
n
[ti=1 O'~ + i""jL PO'iO'j]
1 [ (l-p) L " O'~+P (" =2 L O'i n i=1 i=1 The inequality now follows since var(i);;;: O. Set
~
- IL = Zi and let
)2] .
z be the mean of the Zi' Then
cov(i, Yi) = E[Z(Zi - z)]
1" =E [ Zi'L Zj ] -E(Z2) n i=1 = AO'~ + BO'i - C, where I-p
A=->O, n
P
n
B=- L O'j' n j=1
C=var(i»O.
Hence if 0'1 = 0'2 = ... = 0'", then cov(i, Yi) = O. Conversely, if cov(i, Yi)=O, then AO'~+BO'i-C=O. Since A>O and C>O, there is one positive root O'i = 0' say. But A, B and C are independent of i, so that the conditions cov(i, Yi) = 0 give the equations AO'~+ BO'i - C = 0
for i = 1, 2, ... , n,
which have the solution 0'1 = 0'2 = ... = 0'".
135 UL (1969). Let G(Oh O2 , ( 3 ) denote the joint probability-generating function of Xl> X 2 and X 3 • (i) P(X2 = 1) = coefficient of O2 in G(I, O2 , 1). (ii) P(X1 = 1, X 2 = 1) = coefficient of 0 1 02 in G(Ol, O2 • 1), whence
P(X1 = 11 X 2 = 1) = P(X1 = 1, X 2 = 1)/P(X2 = 1)
42~
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
(iii) P(X3 = 0) = G(1, 1,0). The probability-generating E(Z) = [G'(6, 6, 6)]9=1'
function
of
Z
is
G(6, 6, 6)
an«
136 UL (1969). E(e tX ) = e lLt +!o-2,
and
E(e ut ,+vt 2 ) = E[exp[(t1 + t2)x1 + (t 2- t1)X2]]
= eo-2t~ . exp[2/Lt2+ u2t~], whence the result by the Inversion Theorem.
J 1 J =u-/; e00
E(lui)=_12u-/;
lule- u2/ 4
O-
2
du
-00
00
u
u 2/4cr2
2u du = -/;.
o
If Xl> X 2 , X3 are three independent observations of X, then
IX2 - XII + IX3- X 2 1+ IX3- XII = 2R, where for i l' j E(IXi
-
Xj D= 2u/.Jrr.
137 UL (1969). Let X, Y and Z be the events that A, Band C respectively have success in a game. Then for A to win, the eight possible outcomes of the fin game may be classified as follows. E I : XYZ, XYZ-game reverts to initial state P(E I )
=
p3+ q3= 1-3pq
E 2 : XYZ, XYZ, XYZ-A loses P(E2)
=
pq(1 +q)
E3: XYZ-A wins; P(E3) = pq2 E 4 : XYZ-game undecided between A and C; P(E4 ) = p2q Es: XYZ-game undecided between A and B; P(Es) = p2q. Next, if S.. denotes the event that A wins the series on the nth game, the for n~2 5
P(S..) =
L P(S.. I Ei )P(E;) i=1
where, by definition, P(S.. I E 1) = Un-I;
P(S.. I E 2 ) = P(S.. I E 3 ) = 0;
P(S.. I E 4 ) = P(S.. I Es) = (1- 2pq) .. -2pq. Hence the stated difference equation for 00
L n=2
Z.. -1u., = (1-3pq)
u.,.
It then follows that
00
L n=2
00
Z ..
-IUn_1 +2 p3q 2
L (1-2pq) .. n=2
-2
z n-1
430
EXERCISES IN PROBABILITY AND STATISTICS
or
1 2 p 3q 2 z 2 -
138 UL (1969). When A has £n another trial of E must be made. If this trial results in the occurrence of E, then the expected number of further trials needed for A's win is u,.+l, whereas if this trial results in the non-occurrence of E, then the expected number of further trials needed for A's win is Un-l. Hence Un = 1 + Pu,.+l + qU,.-l,
where Uo = 0 and Ua+b = O.
The difference equation is reduced to the homogeneous form Vn = pVn+l + qVn-l by the substitution u,. = Vn +an where a = 1!(q - p). Hence Un =_n_+ kl(q!p)n + k2 q-p where, by using the boundary conditions the constants (a+b)pa+b k l =-k2=-( q-p )( q a+b -p a+b)' (P=l=q)· The expected duration of play for A to win is Ua • When p = q =!, the difference equation for u,. becomes 2u,. = 2+ U,.+l + Un-l· This is reduced to the homogeneous form 2wn = Wn+l + Wn-l by the substitution u,. = Wn + (3n2, where (3 = -1. Hence the solution is now un =(a+b)n-n 2 and ua=ab.
139 UL (1969). Clearly P(Zj
=
n
0) = 1-N .
Hence E(z.) =!!:... var(z.) = n(N - n) . I N' I N2' 1 N _ (i) E(x) =- L XjE(zj) = X nj=l (ii) var(x) =
:2 [tl XT
var(Zj)+
j~j Xj~ cov(Zi> Zj) J.
which reduces to the given answer.
140 UL (1969). By using the multinomial distribution, the probability density function of
431
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
the joint distribution of Xr and Xs is N! (r-1)! (s-r-1)! (N-s)! I! I! [F(xr)]r-l[F(xs)-F(xr)]S-r-l[I-F(xs )]N-S dF(x ) dF(x )
s x -r- - dXr
dxs
For the uniform distribution F(x) = x/a, whence the distribution of u and v. For fixed u, u:s;; v:s;; 1, so that the probability density function of the marginal distribution of u is 1
N! ------:------:----:--,(r-l)! (s - r-l)! (N -s)!
J r-l( U
V - U
)s-r-l(1 - v ) d v = --:-----'-----...:...-u r- 1(1- u)N-r B(r, N - r+ 1)
" by using the substitution v - u = z(l- u). The conditional distribution of v given u has the probability density function B(N - s + 1, s - r)(I- u)N-r , P(v > vo> u) is evaluated by using the substitution v - u = z(l- u) for
integration over v.
141 UL (1969). x
00
P(X~ Y) = JJnl n2 e-(n,x+n 2 )dy dx Y
o
0
n2 nl +n2
.
.
- - - on mtegratIOn. Hence for X ~ Y, the probability density function of the conditional joint distribution of X and Y is nl(nl + n2) exp- (nix + n2Y) = nl(nl + n2) exp-[nl(x - y)+ (nl + n2)Y], X~
Y;
O:s;; Y <00.
lf we now set u = x - Y and v = y, then it readily follows that U and V are
independently distributed. Hence, for U ~ 0, the conditional distribution of U has the density function 00
J
nl(nl + n2) e- n," e-(n,+n2)u dv = nl e-n,u. o
Again, by symmetry the conditional distribution of Y - X given Y - X ~ 0 has the density function n2 e- n2". Also, P(X ~ Y) = n2/(nl + n2) and P(X:s;; y) = nl/(nl + n2)' Hence the unconditional distribution of IUI has the density function
432
EXERCISES IN PROBABILITY AND STATISTICS
142 UL (1969). Let Xi be a random variable such that Xi = 1 if the ith quadrat is empty = 0 if the ith quadrat is not empty. Then
and 11
X=
L X;.
i=1 It is now easily shown from first principles that
cov(x;, Xj) =
(1-~)" _(1-~) 2.
Hence the stated expressions for E(X) and var(X). Under the given limiting conditions E(X)- n e->';
var(X) - n e->'(1-e:->').
143 UL (1969). Observe that whether X +1 >0 or I1
X,,+1
<0,
1 ( 1-7 X~+I) .
P(X~x..+l)P(X.;;;x..+l)=4
Hence the probability density function of X,.+1 is (2n + 1)! 1 ( X~+I)" n!n!1!(2a)·2211 1-7 '
-a';;;x..+l';;;a.
Therefore
J 1
r = (2n + 1)! a (n !f2211 + 1
o
t(r-1)12(1- t)" dt where X~+1 = t ' a2
whence the stated result.
144 UL (1969). 1 [" II (i) var(x) = n2 i~l var(x;)+ i~ j~i cov(X;, Xj) ] 1
["-1
n-2
]
= n2 nu 2+2 i~ cov(X;, X;+1)+2 i~ cov(X;, Xi+2) ,
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
433
whence the result on reduction. (ii)
E[~ (Xj-X)2]=E[~ XT-nx 2] =
n var(Xj) - n var(x), since E(Xj) = E(x) = 0,
whence the result on reduction. (iii) Since var(x) ~ 0, it follows that p ~ -n/2(2n - 3) for n ~ 2. Hence as n~oo, p~-l.
145 UL (1969). The Jacobian of the x ~ u transformation is n! Also, observe that n
UI = n(xl-m
and
n
L Uj = L (Xi -Xl)'
i=2 i=2 It then follows that the Uj are independent and identically distributed random variables such that for any j"2u/a has the X2 distribution with 2 d.f. The ratio n(n -l)(XI - ml'f.r=2 (Xj - Xl) has the F distribution with 2, 2(n-1) d.f.
146 UL (1966). Clearly, both 20u and 2v/0 are independently distributed as X2'S each with 2n d.f., whence the joint distribution of u and v. Next, the Jacobian of the (u, v) ~ (w, z) transformation is 2w/z. Hence the probability density function of the joint distribution is 2 {f(nWexp[-(Ow/z+wz/0)].w2n-l.z-t, O:os;w
E(z)=
2
B(n, n)
f(Oz-l+zO-I)-2n dz, o
and the integral is readily evaluated 'I'J = 0 2 (Z2+ ( 2), where O:OS;'I'J:OS; 1.
by
using the substitution
147 UL (1966). The joint probability density function of the distribution of Xl and Xn is
,
1! 1! ~~-2)!
[fxf(x) dx ]n-2f(XI)f(Xn),
-00< Xl <00; XI:OS;Xn <00.
Also, given Xl and x'" the probability that the r observations of the second sample all lie outside the range (xt> Xn) is
Hence the integral expression for P. To evaluate the double integral use the
434
EXERCISES IN PROBABILITY AND STATISTICS
transformation
",
J
u=
v=
f(x) dx,
r
f(x) dx.
Then 1
1
JJ
P = n(n -1) du (v - u)"-2[1- (v - u)]r dv. o
lC
The integration over v is carried out by using the substitution z = v - u. To obtain the approximation for P, note that for large n
(n:!r)!=n-r(1-~)(1-r:1) ... (1-~)
[1-'!' i t] n-r[ 1- r(;: 1)].
- n- r =
n
1=1
148 UL (1967). co
E(8 X ) =
e- A
L (A8)'/r! = eMII-n. r=O
Then G(81) ( 2)=E(8i'. 8n = E([8182)X,] . E(8~2) = eXp[Al(8 1 82-1)+ A2(82 -1)]. The probability-generating function of Y is G(1, ( 2) whence P(Y = j) = e-(A ,+A 2>(A) + A2)i/j!, 0 ~ Y ~oo. The coefficient of 8~ in G(81) ( 2) is e-(A ,+A 2>(A2 + AI ( 1)i/ j!, so that the probability-generating function of the conditional distribution of Xl given Y= j is G(8 1I Y = j) = e-(A ,+A 2>(A2+.~181)i/P(Y = j) J. = (1 + p( 1)i(1 + p)i On expansion, the coefficient of 8 1 gives P(X1 = r I Y = j).
149 UL (1966). Unconditionally, 0 ~ X ~ a/2 and 0 ~ Y ~ a. For fixed X, X ~ Y ~ a - X. Hence on integrating the joint distribution with respect to Y, the probability density function of the marginal distribution of X is n(n -1)a-"2"-lx"-2(a -2x),
O~X ~a/2.
For Y in the interval 0 ~ Y ~ a/2, 0 ~ X ~ Y and for Y in the interval
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
al2::o.;; y::o.;; a, 0::0.;; X ::0.;; a - Y. Hence, on integration
435
h
forms of the probability density function of Yare t e corresponding two (i) na-"2,,-1 y"-I, 0::0.;; y::o.;; al2 (ii) na-"2,,-I(a - y)"-\ al2::o.;; y::o.;; a. The probability density function of the conditional distrib t' X=x is (a-2x)-\ for x::o.;;Y::o.;;a-x. Hence U Ion of Y givl'lI (a - xy+l- x r + 1 E(yrIX=x)= (a-2x)(r+1) ,
r;::O
whence E(Y I X = x) and var(Y I X = x).
150 UL (1966). ~x)=t+"'.(x) a~d (x)[1-(x)]=!-",2(x). On expansion of the mtegrand m the detinmg mtegral for "'(x) and term-wise integration . obtain ' we
and
Hence 2X2 2X4 7x 6 4<1>(x)[1-(x)]-1-- + - - - + ...
7T
-e
37T
457T
-2X2hr[1 + 2( 7T - 3) 37T 2
x 4 -'"
]
.
151 UL (1967). Here P(Y ::o.;;ta) = 1- P(Y~ta) -1-
'f' ' "'f-x o
~a
dy dx a(a - x)
=t(1-log 2) =P(a - X - y::o.;;ta),
by symmetry.
After the first experiment there are two possibilities: (i) probability p that Xo does not split; (ij) probability 1- p that Xo splits into two particles XI and X 2 with energies EI and Eo-EI respectively. Now P(E I ::0.;; tEo) =t. Also, if EI ::0.;; tEo, then any further splitting or not of XI must lead to particles or particle with energy 1!Eo. Therefore the probability of Xl giving particles with energy ::o.;;tEo after the second experiment is t(1- p). Next, the probability that X 2 has energy ~tEo is ! and the probability that in the second experiment X 2 splits into two further particles X3 and X 4 is 1- p. If the energies of these particles are E2 and Eo - EI - E 2 , then it is required that E 2 ::o.;;tEo and E o -E I -E2 ::o.;;tEo simultaneously.
436
EXERCISES IN PROBABILITY AND STATISTICS
Hence the total probability required is
!(1-p)+!(1- P)2[P(E2 ~!Eo) xP(Eo- El - E2~!Eo)] where P(E2~!Eo) =P(Eo-
E 1 - E2~!Eo) =!(1 + log 2).
152 UL (1967). 00
E(eIX)=~
J
exp[tx-\x-O\]dx
00
=~
J
exp[t(u+O)-\u\]du 00
=! e 18
00
[J e-(I+I)u du + Je-(I-I)u dU]' o
=
\t\ < 1
0
e '8 (1- t 2)-I,
whence E(X) = 0; var(X) = 2; E(i) = 0; var(i) = 2/n. The moment-generating function of i is
( t2)-"
Mx(t)=E(eIX)=e'8 1- n 2
,
whence
t 2 )-" . -_ ( 1- 2n
Hence, for large n, E(e 'Z ) ~ ell'. For the approximating distribution the proportionality factor is 1/[.J2;(1 + a + 313)] and
( ) _1+3a+1513. varz1 +a+ 3 13 ' ( ) _ 3+15a + 10513 fL4 z 1+a+313 .
E(z) = 0;
For the exact distribution of z, E(z) = 0;
var(z) = 1;
Kiz) =3/n.
Hence
6
a =-613=---. 8n+3
153 UL (1967). If y denotes the number of points obtained uppermost on anyone die, then 6
6
E(y) =
L i.s=!; 0=1
E(y2) =
L i. S2 =9j; 0=1
var(y) =~~.
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
Hence
7nl E(XI)=T;
7 E(X2) = E(X3) = 2" n2.
35nl var ( Xl) =1:2 ;
35n2 var(x2) = var(x3) = 1:2 .
437
It then follows that
var(XI + X2) = var(xi + X3) = H(nl + n2); COV(XI + X2, Xl + X3) = E[(XI + X2)(X2 + X3)]- E(XI + x2)E(XI + X3) = E(x~) + E(XI)E(X2) + E(XI)E(X3) + E(X2)E(X3) -
?(nl + n2)2
=~~nl'
Hence corr(XI + X2, Xl + X3) = ~ . nl+n2
Given
X2 + X3
= c, then
X3
var(xi +x31
= c-
X2'
X2+X3 =
It then follows that
c) =var(xI- X2+C)
= var(xI) + var(x2)' Similarly, it may be shown that COV(XI + X2, Xl + x31
X2 + X3 =
c) = ~~(nl -
n2)'
In the special cases (i), (ii), (iii), corr(xi + X2' Xl + X3) =!. 1,0 and corr(XI + X2, Xl + x31 X2 + X3 = c) = 0, 1, -1.
154 UL (1967). The probability density function of X is k(1 + X 4 )-1 where k = ..fi/7I". The probability density function of Z is 1 _1 _:I - - Z 4(1- z) 4 for O:os;; Z:OS;; 1. 7I".fi. ' The probability density function of the joint distribution of Wand Y is 2 Iyl 71"2 . (1 + w4y4)(1 + y4) whence the probability density function of W is
438
EXERCISES IN PROBABILITY AND STATISTICS
If X and Yare independent N(O, 1) variables, then the joint distribution of Y and W = XI Y has the probability density function
1
27T exp[ _!y2(1 + w2 )] .\y\
whence the probability density function of W is ~
2~
!J ~
Jexp[_!y2(1 + w
2 )]
\y\ dy
=
exp[ _!y2(1 + w2 )] • y dy
o
1
155 UL (1967). The distribution of x,. has the probability density function
I (X2)"-l 2x
n (11-1)! I!
o~'
0;
whence for any r> 0
J
'" E(xr) =2n - X211 + r- l dx 11 0211 11 11
(
2n = -) or. 2n + r
o
It then follows that n0 2
var(xl1 ) = (11 + 1)(211 + 1)2· Hence & and var(&) as stated. Note that the limits for z are
y
-2[n(11 + 1)]~"'; z"'; (n: 1
and the probability density function of z is
J
211-1 2n ( 2110 )211-1 [ Z no 1 211 0 2n + 1 1 + 2{n(n + I)}! . 2n + 1 . [11(n + I)]!
J
2 Z 1 )-~ --i>e'-1 1 -)211[ 1+ (1 - • 11-1 ( 1+2n + 1 2{n(n + I)}> n for -00< z < 1. =
as n --i> 00,
156 UL (1967). Probability of a withdrawal only in the week before Christmas is PI - P12; Probability of a withdrawal only in the week after Christmas is P2 - P12; Probability of no withdrawals in these two weeks is 1- PI - P2 + P12. Hence the probability-generating function of the withdrawals of a single client in the two weeks is (1- PI - P2+ pd+(Pl- P12)81 +(P2- pd 02+ PIP2 0 1 02 = 1 + Pl(OI -1) + P2(02-1) + PliOI -1)(02 -1),
whence E(O~'. 0~2)= G(Ol, O2 ).
ANSWERS AND HINTS ON SOLUTIONS: SUPPLEMENT
439
Under the stated conditions the limiting form of G(O 1> 0)' 2 IS exp[al(OI -1) + a2(02 -1)+ aliOI -1)(02 -1)]. The moment-generating function about the origin of Xl and X is G(' , whence the cumulants on logarithmic expansion. For corr(X 2X ) == eO" eh,) . f unction . fX I and 2 ,tc moment-generatIng 0 X 2'IS exp[a l (e"-l)1>+a2(e"-1)] whence independence of Xl and X 2 • '
157 UL (1969). (i) The probability-density function of the marginal distribution of Y is =
(y )
(3-a-2 cxI 1 r( a + 1) . y -(l-+-x-)-=-/3 exp - -l-+-x dx. o
To integrate over x for fixed y use the substitution w = y/(l + x), whence the stated result. (ii) This follows readily by using P(X I Y = y) = P(X = x, Y = y)/ p(Y=y).
I =
•.•
y/3-1
(111) E(XI Y=y)= g(y)
x e- y /(1+x) (l+x)/3 dx.
o
This integral is evaluated by using the substitution z = y/(l + x). The Jacobian of the (X, Y) ~ (U, V) transformation is v- 3 , whence the probability density function of the joint distribution of U and V is
1
---e-uu cx (Q-a-2)v/3-cx-3
r(a + 1)
.
tJ
O,.;;U
,
,
whence independence and marginal distributions.
158 UL (1969). P(X ~ Xl) =
~
I =
e-(x-/3)/cx
dx
= e-(x,-/3)/CX.
The probability distribution of Xl is now obtained by using the multinomial distribution. E(XI) follows directly on integration. Since Xi ~ Xi> therefore for fixed Xl the probability density function of the conditional distribution of Xi is
1
_ e-(x,-/3)/cx/e -(x,-/3)/cx
a
= _1 e-(x,-x,)/cx. a
Hence E(a* I Xl) = a = E(a*). E«(3*) follows on taking expectations.
159 UL (1969).
(i) The limits for ware obtained from I:'=l (xi-i)2~O which may be rewritten as n - W2~O.
440
EXERCISES IN PROBABILITY AND STATISTICS
(ii) Note that
w=
nx [en _1)S2 + nx2]~
=-:------:---;;------;;-::;-
whence the relation between t 2 and w2 • The sign of t is the same as that of w. Also, n - w 2 ;;;. O. Hence as w2 ~ n, t 2 ~ 00. Therefore t is a monotonic function of w. Note that u = wl..fii = vl(1 + v 2 )! and the suggested transformations lead to the stated distribution of v.
160 UL (1969). Here E(Y1) = E(Y2 ) = O. var(Y1) = !( Y 2) = !( Y2 are uncorrelated if P
-I
2
1
2·
For this value of 0, Y1 and Y2 are independent normal variables with zero ' means and variances