Lecture 32
Lecture 32
MODULE 6
RANDOM VECTOR AND ITS JOINT DISTRIBUTION
LECTURE 32
Topics
6.10 DISTRIBUTION OF FUNCTION OF RANDOM
VECTORS
6.10.1 Distribution Function Technique
Example 10.1
Consider a company that manufactures electric bulbs. The lifetimes of electric bulbs
manufactured by the company are random. Past experience with testing on electric bulbs
manufactured by the company suggests that the lifetime of a randomly chosen electric
bulb manufactured by the company can be described by a random variable 𝑋 having the
p.d.f.
1 −𝑥
𝑓𝑋 𝑥|𝜃 = 𝜃 𝑒 , if 𝑥 > 0
𝜃
, 𝜃 > 0.
0, otherwise
However the value of 𝜃 > 0 is not evident from the past experience and therefore 𝜃 is
unknown. One way to obtain information about unknown 𝜃 is to do testing independently
and under identical conditions, on a number (say 𝑛) of electric bulbs manufactured by the
𝑛 𝑛 𝑥
1 − 𝑖=1 𝑖
𝑥𝑖 |𝜃 = 𝜃 𝑛 𝑒 , if 𝑥𝑖 > 0, 𝑖 = 1, … , 𝑛 .
𝜃
𝑓𝑋 𝑥|𝜃 = 𝑓𝑋 𝑖
𝑖=1 0, otherwise
1 𝑛
Since 𝐸 𝑋 = 𝜃, a natural estimator of 𝜃 is the sample mean 𝑋 = 𝑛 𝑖=1 𝑋𝑖 .To study
theoretical properties of the estimator 𝑋 we need probability distribution of 𝑋. ▄
Definition 10.1
(i) A function of one or more random variables that does not depend on any unknown
parameter is called a statistic.
(ii) Let 𝑋1 , … , 𝑋𝑛 be a collection of independent random variables each having the same
p.m.f./p.d.f. 𝑓(or distribution function 𝐹). We then call 𝑋1 , … , 𝑋𝑛 a random sample
(of size 𝑛) from a distribution having p.m.f./p.d.f 𝑓 (or distribution function 𝐹). In
other words a random sample is a collection of independent and identically
distributed random variables.
Remark 10.1
(i) Let 𝑋 ~ 𝑁2 𝜇1 , 𝜇2 , 𝜍12 , 𝜍22 , 𝜌 , −∞ < 𝜇𝑖 < ∞, 𝜍𝑖 , 𝑖 = 1, 2, −1 < 𝜌 < 1. Then the
𝑋1 −𝜇 1
random variable 𝑌1 = 𝑋1 + 𝑋2 is a statistic but the random variable 𝑌2 = is not
𝜍1
a statistic unless 𝜇1 and 𝜍1 are known parameters.
(ii) Although a statistic does not depend upon any unknown parameters, the distribution
of a statistic may very well depend upon unknown parameters. For example, in (i)
above, 𝑌1 ~ 𝑁 𝜇1 + 𝜇2 , 𝜍12 + 𝜍22 + 2𝜌𝜍1 𝜍2 .
(iii) If 𝑋1 , … , 𝑋𝑛 is a random sample from a distribution having p.m.f./p.d..f. 𝑓 ∙ , then the
joint p.m.f./p.d.f. of 𝑋 = (𝑋1 , … , 𝑋𝑛 ) is
𝑛
𝑓𝑋 𝑥1 , … , 𝑥𝑛 = 𝑓𝑋 𝑖 𝑥𝑖
𝑖=1
= 𝑓 𝑥𝑖 , 𝑥 = 𝑥1 , … , 𝑥𝑛 ∈ ℝ𝑛 .
𝑖=1
(iv) Let 𝑋1 , … , 𝑋𝑛 be a random sample from a distribution. Some of the commonly used
statistics are
𝑛
1
a Sample Mean 𝑋 = 𝑋𝑖 ;
𝑛
𝑖=1
𝑛 𝑛
1 1
b Sample Variance 𝑆 = 2
𝑋𝑖 − 𝑋 2
= 𝑋𝑖2 − 𝑛𝑋 2 , 𝑛 ≥ 2 ;
𝑛−1 𝑛−1
𝑖=1 𝑖=1
𝑋𝑛 +1:𝑛 , if 𝑛 is odd
2
e Median 𝑀 = 𝑋𝑛 :𝑛 + 𝑋𝑛 +1:𝑛 .
2 2
, if 𝑛 is even
𝑛
Theorem 10.1
Let 𝑋1 , … , 𝑋𝑛 be a random sample from a distribution having p.m.f./p.d.f. 𝑓(⋅). Then, for
any permutation 𝛽1 , … , 𝛽𝑛 of 1, … , 𝑛 ,
𝑑
𝑋1 , … , 𝑋𝑛 = 𝑋𝛽 1 , … , 𝑋𝛽𝑛 .
𝑓𝑋 𝛽 𝑥1 , … , 𝑥𝑛 = 𝑓𝑋1 ,…,𝑋𝑛 𝑥𝛾 1 , … , 𝑥𝛾 𝑛
1 ,…,𝑋 𝛽 𝑛
= 𝑓𝑋 𝑖 𝑥𝛾 𝑖
𝑖=1
= 𝑓 𝑥𝑖
𝑖=1
= 𝑓𝑋1 ,…,𝑋𝑛 𝑥1 , … , 𝑥𝑛 .
It follows that
𝑑
⇒ 𝑋𝛽1 , … , 𝑋𝛽𝑛 = 𝑋1 , … , 𝑋𝑛 . ▄
Example10.2
(i) If 𝑋1 is of absolutely continuous type then show that 𝑃 𝑋1 < 𝑋2 < ⋯ < 𝑋𝑛 =
1
𝑃 𝑋𝛽 < 𝑋𝛽 < ⋯ < 𝑋𝛽 = 𝑛!, for any permutation (𝛽1 , … , 𝛽𝑛 ) of 1, … , 𝑛 ;
1 2 𝑛
1
(ii) If 𝑋1 is of absolutely continuous type then show that 𝑃 𝑋𝑖 = 𝑋𝑟:𝑛 = 𝑛 , 𝑖 = 1, … , 𝑛,
where, for 𝑟 ∈ 1, … , 𝑛 , 𝑋𝑟:𝑛 = 𝑟-th smallest of 𝑋1 , … , 𝑋𝑛 ;
(iii) Show that
𝑋𝑖 1
𝐸 = , 𝑖 = 1,2, … , 𝑛,
𝑋1 + 𝑋2 + ⋯ + 𝑋𝑛 𝑛
Solution. Let 𝑆𝑛 denote the set of all permutations of 1, … , 𝑛 . Using Theorem 10.1 we
have
𝑑
𝑋1 , … , 𝑋𝑛 = 𝑋𝛽 , … , 𝑋𝛽 , ∀𝛽 = 𝛽1 , … , 𝛽𝑛 ∈ 𝑆𝑛 .
1 𝑛
⇒ 𝐸 Ψ 𝑋1 , … , 𝑋𝑛 = 𝐸 Ψ 𝑋𝛽 , … , 𝑋𝛽 , ∀𝛽 ∈ 𝑆𝑛 (10.1)
1 𝑛
(i) On taking
We conclude that
1
⇒ 𝑃 𝑋1 < 𝑋2 < ⋯ < 𝑋𝑛 =𝑃 𝑋𝛽 < 𝑋𝛽 < ⋯ < 𝑋𝛽 = 𝑛! . (using (10.2))
1 2 𝑛
𝑃 𝑋𝑖 = 𝑋𝑟:𝑛 =𝑃 𝑋𝛽 = 𝑋𝑟:𝑛 ∀ 𝛽 ∈ 𝑆𝑛
𝑖
⇒ 𝑃 𝑋𝑖 = 𝑋𝑟:𝑛 = 𝑃 𝑋1 = 𝑋𝑟:𝑛 𝑖 = 1, … , 𝑛.
But
𝑛
𝑃 𝑋𝑖 = 𝑋𝑟:𝑛 = 1,
𝑖=1
and therefore
1
𝑃 𝑋𝑖 = 𝑋𝑟:𝑛 = 𝑃 𝑋1 = 𝑋𝑟:𝑛 = ∙
𝑛
(iii) On taking
𝑥1
Ψ 𝑥1 , … , 𝑥𝑛 = , 𝑥 ∈ ℝ𝑛 ,
𝑥1 + ⋯ + 𝑥𝑛
𝑋1 𝑋𝛽
1
𝐸 =𝐸
𝑋1 + ⋯ + 𝑋𝑛 𝑋𝛽 + ⋯ + 𝑋𝛽
1 𝑛
𝑛 𝑛
𝑋𝛽
1
=𝐸 since 𝑥𝑖 = 𝑋𝛽
𝑋1 + ⋯ + 𝑋𝑛 𝑖
𝑖=1 𝑖=1
𝑋1 𝑋𝑖
⇒𝐸 =𝐸 , 𝑖 = 1, … , 𝑛. (10.3)
𝑋1 + ⋯ + 𝑋𝑛 𝑋1 + ⋯ + 𝑋𝑛
But
𝑛 𝑛
𝑋𝑖 𝑋𝑖
𝐸 =𝐸
𝑋1 + ⋯ + 𝑋𝑛 𝑋1 + ⋯ + 𝑋𝑛
𝑖=1 𝑖=1
𝑋1 + ⋯ + 𝑋𝑛
=𝐸
𝑋1 + ⋯ + 𝑋𝑛
= 1.
𝑋1 𝑋𝑖 1
𝐸 =𝐸 = , 𝑖 = 1, … , 𝑛.
𝑋1 + ⋯ + 𝑋𝑛 𝑋1 + ⋯ + 𝑋𝑛 𝑛
𝑛 𝑛
⇒ 𝐸 𝑋1 𝑋𝑗 = 𝑡 = 𝐸 𝑋𝛽 𝑋𝑗 = 𝑡 , ∀𝛽 ∈ 𝑆𝑛
1
𝑗 =1 𝑗 =1
𝑛 𝑛
⇒ 𝐸 𝑋1 𝑋𝑗 = 𝑡 = 𝐸 𝑋𝑖 𝑋𝑗 = 𝑡 , 𝑖 = 1, … , 𝑛. (10.4)
𝑗 =1 𝑗 =1
But
𝑛 𝑛 𝑛 𝑛
𝐸 𝑋𝑖 𝑋𝑗 = 𝑡 = 𝐸 𝑋𝑗 𝑋𝑗 = 𝑡 = 𝑡.
𝑖=1 𝑗 =1 𝑖=1 𝑗 =1
𝑛 𝑛
𝑡
𝐸 𝑋1 𝑋𝑗 = 𝑡 = 𝐸 𝑋𝑖 𝑋𝑗 = 𝑡 = , 𝑖 = 1, … , 𝑛. ▄
𝑛
𝑗 =1 𝑗 =1
In the following subsections we will discuss various techniques to find the distribution of
functions of random variables.
𝐹𝑌 𝑦 = 𝑃 𝑔 𝑋1 , … , 𝑋𝑝 ≤ 𝑦 , −∞ < 𝑦 < ∞.
Example 10.1.1
∞
𝑓 𝑥 𝐼𝐶 𝑐 𝑥 𝑑𝑥 = 1.
−∞
𝐹𝑋 𝑥1 , … , 𝑥𝑛 = 𝐹 𝑥𝑖 , 𝑥 ∈ ℝ𝑛 .
𝑖=1
We have
𝑛
𝐹𝑋 𝑥 = 𝐹 𝑥𝑖
𝑖=1
𝑛 𝑥𝑖
= 𝑓 𝑡𝑖 𝑑𝑡𝑖
𝑖=1 −∞
𝑥1 𝑥𝑛 𝑛
= ⋯ 𝑓 𝑡𝑖 𝑑𝑡𝑛 ⋯ 𝑑𝑡1
−∞ −∞ 𝑖=1
𝑥1 𝑥𝑛
= ⋯ 𝑓𝑋 𝑡 𝑑𝑡𝑛 ⋯ 𝑑𝑡1 .
−∞ −∞
It follows that 𝑋 is of absolutely continuous type with joint p.d.f. 𝑓𝑋 (⋅). Therefore, for
𝑖 ≠ 𝑗,
𝑃 𝑋𝑖 = 𝑋𝑗 = 0.
Define
𝑋𝑟:𝑛 = 𝑟 − th smallest of 𝑋1 , … , 𝑋𝑛 , 𝑟 = 1, … , 𝑛,
so that
Therefore
𝐹𝑋 𝑟:𝑛 𝑥 = 𝑃 𝑋𝑟:𝑛 ≤ 𝑥
= 𝑃 at least 𝑟 of 𝑋1 , … , 𝑋𝑛 are ≤ 𝑥
𝑛
= 𝑃 𝑖 of 𝑋1 , … , 𝑋𝑛 are ≤ 𝑥 , 𝑥 ∈ ℝ .
𝑖=𝑟
Fix 𝑥 ∈ ℝ, and consider a sequence of 𝑛 trials where at the 𝑖-th trial we observe 𝑋𝑖 and
consider the trial having resulted in a success if 𝑋𝑖 ≤ 𝑥 and it having resulted in a failure
if 𝑋𝑖 > 𝑥, 𝑖 = 1, … , 𝑛. Since 𝑋1 , … , 𝑋𝑛 are independent and the probability of success in
the 𝑖-th trial is 𝑃 𝑋𝑖 ≤ 𝑥 = 𝐹 𝑥 (same for all the trials), the above sequence of trials
may be considered as a sequence of independent Bernoulli trials with probability of
success in each trial as 𝐹 𝑥 . Therefore
𝑃 𝑖 of 𝑋1 , … , 𝑋𝑛 ≤ 𝑥 = 𝑃 𝑖 successes in 𝑛 trials
𝑛 𝑖 𝑛−𝑖
= 𝐹 𝑥 1−𝐹 𝑥 ,
𝑖
and consequently
𝑛
𝑛 𝑖 𝑛−𝑖
𝐹𝑋 𝑟:𝑛 𝑥 = 𝐹 𝑥 1−𝐹 𝑥 , 𝑥 ∈ ℝ.
𝑖
𝑖=𝑟
Therefore,
𝐹(𝑥)
1
𝐹𝑋 𝑟:𝑛 𝑥 = 𝑡 𝑟 −1 1 − 𝑡 𝑛 −𝑟
𝑑𝑡, 𝑥 ∈ ℝ .
𝐵 𝑟, 𝑛 − 𝑟 + 1
0
Let
1 𝑟−1 𝑛−𝑟
𝑓𝑋 𝑟:𝑛 𝑥 = 𝐹 𝑥 1−𝐹 𝑥 𝑓 𝑥 , 𝑥 ∈ ℝ, (10.1.1)
𝐵 𝑟, 𝑛 − 𝑟 + 1
so that
𝑑
𝐹 𝑥 = 𝑓𝑋 𝑟:𝑛 𝑥 , ∀𝑥 ∉ 𝐶 𝑐 ,
𝑑𝑥 𝑋 𝑟:𝑛
and
∞ ∞
𝑓𝑋 𝑟:𝑛 𝑥 𝐼𝐶 𝑐 (𝑥)𝑑𝑥 = 𝑓𝑋 𝑟:𝑛 𝑥 𝐼𝐶 𝑐 (𝑥)𝑑𝑥
−∞ −∞
∞
1 𝑟−1 𝑛−𝑟
= 𝐹 𝑥 1−𝐹 𝑥 𝑓 𝑥 𝐼𝐶 𝑐 (𝑥)𝑑𝑥
−∞ 𝐵 𝑟, 𝑛 − 𝑟 + 1
1
1
= 𝑡 𝑟 −1 1 − 𝑡 𝑛−𝑟
𝑑𝑡
𝐵 𝑟, 𝑛 − 𝑟 + 1 0
= 1.
Using Remark 4.2 (vii), Module 2, it follows that the random variable 𝑋𝑟:𝑛 is of
absolutely continuous type with p.d.f. given by (10.1.1). A simple heuristic argument for
expression (10.1.1) is as follows. Interpret 𝑓𝑋 𝑟:𝑛 𝑥 ∆𝑥 as the probability that 𝑋𝑟:𝑛 lies in
an infinitesimal interval 𝑥, 𝑥 + ∆𝑥 . Realizing that the probability of more than one 𝑋𝑖′ 𝑠
falling in the infinitesimal interval 𝑥, 𝑥 + ∆𝑥 may be negligible, 𝑓𝑋 𝑟:𝑛 𝑥 ∆𝑥 may be
interpreted as probability that one of the 𝑋𝑖′ 𝑠 falls in the infinitesimal interval 𝑥, 𝑥 +
∆𝑥 , 𝑟 − 1 𝑋𝑖′ 𝑠 fall in the interval (−∞, 𝑥] and 𝑛 − 𝑟 𝑋𝑖′ 𝑠 fall in the interval (𝑥 +
∆𝑥, ∞) ≃ 𝑥, ∞ . Since 𝑋1 , … , 𝑋𝑛 are independent and the probabilities of an observation
falling in intervals 𝑥, 𝑥 + ∆𝑥 , −∞, 𝑥 and 𝑥, ∞ are 𝑓 𝑥 ∆𝑥, 𝐹 𝑥 and 1 − 𝐹 𝑥
respectively, 𝑓𝑋𝑟:𝑛 𝑥 ∆𝑥 is given by the multinomial probability
𝑛! 1 𝑟−1 𝑛−𝑟
𝑓𝑋𝑟:𝑛 𝑥 ∆𝑥 ≡ 𝑓 𝑥 ∆𝑥 𝐹 𝑥 1−𝐹 𝑥 ,
1! (𝑟 − 1)! (𝑛 − 𝑟)!
i.e.,
𝑛! 𝑟−1 𝑛−𝑟
𝑓𝑋𝑟:𝑛 𝑥 = 𝐹 𝑥 1−𝐹 𝑥 𝑓 𝑥 , −∞ < 𝑥 < ∞.
(𝑟 − 1)! (𝑛 − 𝑟)!
Now we will derive the joint distribution of 𝑋𝑟:𝑛 , 𝑋𝑠:𝑛 , where 𝑟 and 𝑠 are fixed positive
integers satisfying 1 ≤ 𝑟 < 𝑠 ≤ 𝑛. For −∞ < 𝑥 < 𝑦 < ∞,
𝐹𝑋𝑟:𝑛 ,𝑋𝑠:𝑛 𝑥, 𝑦
= 𝑃 𝑋𝑟:𝑛 ≤ 𝑥, 𝑋𝑠:𝑛 ≤ 𝑦
𝑛! 𝑖 𝑗 𝑛−𝑖−𝑗
= 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥 1−𝐹 𝑦 ∙
𝑖! 𝑗! (𝑛 − 𝑖 − 𝑗)!
𝑛 𝑛 −𝑖
𝑛! 𝑖 𝑗 𝑛−𝑖−𝑗
= 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥 1−𝐹 𝑦
𝑖! 𝑗! (𝑛 − 𝑖 − 𝑗)!
𝑖=𝑟 𝑗 =max 0,𝑠−𝑖
𝑠−1 𝑛 −𝑖
𝑛! 𝑖 𝑗 𝑛−𝑖−𝑗
= 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥 1−𝐹 𝑦
𝑖! 𝑗! (𝑛 − 𝑖 − 𝑗)!
𝑖=𝑟 𝑗 =𝑠−𝑖
𝑛 𝑛 −𝑖
𝑛! 𝑖 𝑗 𝑛−𝑖−𝑗
+ 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥 1−𝐹 𝑦
𝑖! 𝑗! (𝑛 − 𝑖 − 𝑗)!
𝑖=𝑠 𝑗 =0
𝑠−1 𝑛 −𝑖
𝑛 𝑖 𝑛−𝑖 𝑗 𝑛−𝑖−𝑗
= 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥 1−𝐹 𝑦
𝑖 𝑗
𝑖=𝑟 𝑗 =𝑠−𝑖
𝑛 𝑛 −𝑖
𝑛 𝑖 𝑛−𝑖 𝑗 𝑛−𝑖−𝑗
+ 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥 1−𝐹 𝑦
𝑖 𝑗
𝑖=𝑠 𝑗 =0
𝐹 𝑦 −𝐹 𝑥
𝑠−1 1−𝐹 𝑥
𝑛 𝑖 𝑛−𝑖
1
= 𝐹 𝑥 1−𝐹 𝑥 𝑡 𝑠−𝑖−1 1 − 𝑡 𝑛−𝑠
𝑑𝑡
𝑖 𝐵 𝑠 − 𝑖, 𝑛 − 𝑠 + 1
𝑖=𝑟 0
𝐹 𝑥
1
+ 𝑡 𝑠−1 1 − 𝑡 𝑛 −𝑠
𝑑𝑡. (using Theorem 3.1, Module 5)
𝐵 𝑠, 𝑛 − 𝑠 + 1
0
𝑛−𝑠
𝐹 𝑦 −𝐹 𝑥 𝑓 𝑦
× 1−
1−𝐹 𝑥 [1 − 𝐹 𝑥 ]
𝑛−𝑠 𝑠−1
𝑛! 1 − 𝐹 𝑦 𝑠−1 𝑖 𝑠−𝑖−1
= 𝑓 𝑦 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥
(𝑠 − 1)! (𝑛 − 𝑠)! 𝑖
𝑖=𝑟
𝑛! 𝑠−1 𝑛−𝑠
= 𝐹 𝑦 1−𝐹 𝑦 𝑓 𝑦
𝑠−1 ! 𝑛−𝑠 !
𝑠−1 𝑖 𝑠−𝑖−1
𝑠−1 𝐹 𝑥 𝐹 𝑥
× 1−
𝑖 𝐹 𝑦 𝐹 𝑦
𝑖=𝑟
𝑛! 𝑠−1 𝑛−𝑠
= 𝐹 𝑦 1−𝐹 𝑦 𝑓 𝑦
𝑠−1 ! 𝑛−𝑠 !
𝐹 𝑥
𝐹 𝑦
1
× 𝑡 𝑟 −1 1 − 𝑡 𝑠−𝑟−1
𝑑𝑡
𝐵 𝑟, 𝑠 − 𝑟
0
𝜕2
⇒ 𝑓𝑋𝑟:𝑛 ,𝑋𝑠:𝑛 𝑥, 𝑦 = 𝐹
𝜕𝑥𝜕𝑦 𝑋𝑟:𝑛 ,𝑋𝑠:𝑛
𝑛! 𝑠−1 𝑛−𝑠
= 𝐹 𝑦 1−𝐹 𝑦 𝑓 𝑦
𝑟−1 ! 𝑠−𝑟−1 ! 𝑛−𝑠 !
𝑟−1 𝑠−𝑟−1
𝐹 𝑥 𝐹 𝑥 𝑓 𝑥
× 1−
𝐹 𝑦 𝐹 𝑦 𝐹 𝑦
𝑛! 𝑟−1 𝑠−𝑟−1
= 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥
𝑟−1 ! 𝑠−𝑟−1 ! 𝑛−𝑠 !
𝑛−𝑠
× 1−𝐹 𝑦 𝑓 𝑥 𝑓 𝑦 , −∞ < 𝑥 < 𝑦 < ∞, 𝑥, 𝑦 ∈ 𝐶 𝑐 .
𝑋𝑠:𝑛 ≤ 𝑦 ⊆ 𝑋𝑟:𝑛 ≤ 𝑥
and therefore
𝜕2
⇒ 𝐹 𝑥, 𝑦 = 0, −∞ < 𝑥 < 𝑦 < ∞, 𝑥, 𝑦 ∈ 𝐶 𝑐 ..
𝜕𝑥𝜕𝑦 𝑋𝑟:𝑛 ,𝑋𝑠:𝑛
Let
𝑛! 𝑟−1 𝑠−𝑟−1 𝑛−𝑠
𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥 1−𝐹 𝑦 𝑓 𝑥 𝑓 𝑦 ,
𝑟−1 ! 𝑠−𝑟−1 ! 𝑛−𝑠 !
𝑓𝑟,𝑠 𝑥, 𝑦 = if − ∞ < 𝑥 < 𝑦 < ∞
0, otherwise
(10.1.2)
so that
𝜕2
𝐹 𝑥, 𝑦 = 𝑓𝑟,𝑠 𝑥, 𝑦 ∀ 𝑥, 𝑦 ∈ ℝ2 − 𝐶 × 𝐶 = 𝐷 (say).
𝜕𝑥𝜕𝑦 𝑋𝑟:𝑛 ,𝑋𝑠:𝑛
Using Remark 2.1 (xiii), it follows that the random vector (𝑋𝑟:𝑛 , 𝑋𝑠:𝑛 ) is of absolutely
continuous type with joint p.d.f. given by (10.1.2). One can give the following heuristic
argument for the expression (10.1.2). For −∞ < 𝑥 < 𝑦 < ∞,
𝑓𝑟,𝑠 𝑥, 𝑦 ∆𝑥∆𝑦 = probability that 𝑟 − 1 𝑋𝑖′ 𝑠 fall in (−∞, 𝑥], one 𝑋𝑖 falls in 𝑥, 𝑥 +
Δ𝑥 , 𝑠 − 𝑟 − 1 𝑋𝑖′ 𝑠 fall in 𝑥 + ∆𝑥, 𝑦 ≈ 𝑥, 𝑦 , one 𝑋𝑖 falls in (𝑦, 𝑦 + Δ𝑦] and
𝑛 − 𝑠 𝑋𝑖 s fall in 𝑦 + ∆𝑦, ∞ ≈ 𝑦, ∞ . Using the property of multinomial
distribution, we have
𝑓𝑟,𝑠 𝑥, 𝑦 ∆𝑥∆𝑦
𝑛! 𝑟−1 1 𝑠−𝑟−1 1 𝑛−𝑠
= 𝐹 𝑥 𝑓 𝑥 ∆𝑥 𝐹 𝑦 −𝐹 𝑥 𝑓 𝑦 ∆𝑦 1−𝐹 𝑦
(𝑟 − 1)! 1! (𝑠 − 𝑟 − 1)! 1! (𝑛 − 𝑠)!
i.e.,
𝑛! 𝑟−1 𝑠−𝑟−1
𝑓𝑟,𝑠 𝑥, 𝑦 = 𝐹 𝑥 𝐹 𝑦 −𝐹 𝑥
𝑟−1 ! 𝑠−𝑟−1 ! 𝑛−𝑠 !
𝑛−𝑠
× 1−𝐹 𝑦 𝑓 𝑥 𝑓 𝑦 , −∞ < 𝑥 < 𝑦 < ∞. ▄
Example 10.1.2
𝐹𝑋1:𝑛 𝑥 = 𝑃 𝑋1:𝑛 ≤ 𝑥
= 1 − 𝑃 𝑋1:𝑛 > 𝑥
= 1 − 𝑃 𝑋𝑖 > 𝑥, 𝑖 = 1, … , 𝑛
𝑛
= 1− 𝑃 𝑋𝑖 > 𝑥
𝑖=1
=1− 1−𝐹 𝑥
𝑖=1
𝑛
= 1− 1−𝐹 𝑥 .
Note that
= 𝑥 ∈ ℝ: 𝐹 ∙ is discontinuous at 𝑥
= 𝑆.
Thus 𝑋1:𝑛 is a discrete type random variable with support 𝑆 and p.m.f.
𝐹𝑋1:𝑛 𝑥 − 𝐹𝑋1:𝑛 𝑥 − , if 𝑥 ∈ 𝑆
𝑓𝑋 1:𝑛 𝑥 =
0, otherwise
𝑛 𝑛
1−𝐹 𝑥− − 1−𝐹 𝑥 , if 𝑥 ∈ 𝑆
= .
0, otherwise
𝐹𝑋 1:𝑛 𝑥 = 𝑃 𝑋𝑛:𝑛 ≤ 𝑥
= 𝑃 𝑋𝑖 ≤ 𝑥, 𝑖 = 1, … , 𝑛
𝑛
= 𝑃 𝑋𝑖 ≤ 𝑥
𝑖=1
= 𝐹 𝑥
𝑖=1
𝑛
= 𝐹 𝑥 , 𝑥 ∈ ℝ.
Since 𝐹𝑋𝑛 :𝑛 ∙ is continuous at 𝑥 if, and only if, 𝐹 ∙ is continuous at 𝑥 , the random
variable 𝑋𝑛:𝑛 is of discrete type with support 𝑆 and p.m.f.
𝐹𝑋𝑛 :𝑛 𝑥 − 𝐹𝑋𝑛 :𝑛 𝑥 − , if 𝑥 ∈ 𝑆
𝑓𝑋𝑛 :𝑛 𝑥 =
0, otherwise
𝑛 𝑛
𝐹 𝑥 − 𝐹 𝑥− , if 𝑥 ∈ 𝑆
= .▄
0, otherwise
Example 10.1.3
Let 𝑋1 , 𝑋2 be a random sample from 𝑈 0,1 distribution. Find the distribution function of
𝑌 = 𝑋1 + 𝑋2 . Hence find the p.d.f. of 𝑌.
𝐹𝑌 𝑥 = 𝑃 𝑋1 + 𝑋2 ≤ 𝑥
∞ ∞
1 1
= 𝑑𝑥1 𝑑𝑥2
0 0
𝑥 1 + 𝑥 2 ≤𝑥
0, if 𝑥 < 0
1
× 𝑥 × 𝑥, if 0 ≤ 𝑥 < 1
= 2 .
1
1− 2−𝑥 × 2−𝑥 , if 1 ≤ 𝑥 < 2
2
1, if 𝑥 ≥ 2
0, if 𝑥 < 0
𝑥2
, if 0 ≤ 𝑥 < 1
⇒ 𝐹𝑌 𝑥 = 2 .
4𝑥 − 𝑥 2 − 2
, if 1 ≤ 𝑥 < 2
2
1, if 𝑥 ≥ 2
𝑥, if 0 < 𝑥 < 1
𝑔 𝑥 = 2 − 𝑥, if 1 < 𝑥 < 2 ,
0, otherwise
so that
𝑑
𝐹 𝑥 = 𝑔 𝑥 , ∀𝑥 ∈ ℝ − 𝐶
𝑑𝑥 𝑌
and
∞
𝑔 𝑥 𝑑𝑥 = 1.
−∞
𝑥, if 0 < 𝑥 < 1
𝑔 𝑥 = 2 − 𝑥, if 1 < 𝑥 < 2 . ▄
0, otherwise
Example 10.1.4
𝑓𝑋1 ,𝑋2 𝑥1 , 𝑥2 = 𝑓𝑋 1 𝑥1 𝑓𝑋 2 𝑥2
For 0 ≤ 𝑥 < 1
𝑥 𝑥−𝑥 1
𝑥4
𝐹𝑌 𝑥 = 4𝑥1 𝑥2 𝑑𝑥2 𝑑𝑥1 = .
6
0 0
For 1 ≤ 𝑥 < 2
𝑥−1 1 1 𝑥−𝑥 1
3
2
4𝑥 − 3 − 𝑥 + 3 𝑥 − 1
= 𝑥−1 + .
6
Therefore,
0, if 𝑥 < 0
𝑥4
, if 0 ≤ 𝑥 < 1
𝐹𝑌 𝑥 = 6 .
2
4𝑥 − 3 − 𝑥 + 3 𝑥 − 1 3
𝑥−1 + , if 1 ≤ 𝑥 < 2
6
1, if 𝑥 ≥ 2
2 3
𝑥 , if 0 < 𝑥 < 1
3
𝑔 𝑥 = 2 2
,
2 𝑥−1 + 1− 𝑥+2 𝑥−1 , if 1 < 𝑥 < 2
3
0, otherwise
so that
𝑑
𝐹 𝑥 = 𝑔 𝑥 , ∀𝑥 ∈ ℝ − 𝐶
𝑑𝑥 𝑌
and
∞
𝑔 𝑥 𝑑𝑥 = 1.
−∞
2 3
𝑥 , if 0 < 𝑥 < 1
3
𝑔 𝑥 = 2 2
.▄
2 𝑥−1 + 1− 𝑥+2 𝑥−1 , if 1 < 𝑥 < 2
3
0, otherwise
Example 10.1.5
𝑓𝑋 𝑥1 , 𝑥2 , 𝑥3 = 𝑓𝑋 𝑖 𝑥𝑖
𝑖=1
3
1 𝑥2
𝑖
= 𝑒− 2
𝑖=1
2𝜋
1 −
1
𝑥 12 +𝑥 22 +𝑥 32
= 3𝑒
2 , − ∞ < 𝑥𝑖 < ∞, 𝑖 = 1, 2, 3.
2𝜋 2
𝑥1 = 𝑟 sin 𝜃1 sin 𝜃2 ,
𝑥2 = 𝑟 sin 𝜃1 cos 𝜃2 ,
𝑥3 = 𝑟 cos 𝜃1 ,
𝑦
2 𝑟2
= 𝑒− 2 𝑟 2 𝑑𝑟
𝜋
0
𝑦
1 𝑡 3
= 3 𝑒 − 2 𝑡 2− 1 𝑑𝑡
3
22 Γ(2) 0
Therefore
0, if 𝑦 ≤ 0
𝑦
𝐹𝑌 𝑦 = 1 𝑡 3
.
3 𝑒 −2 𝑡 2 − 1 𝑑𝑡, if 𝑦 > 0
3
2 Γ(2)
2
0
𝐹𝑌 𝑦 = 𝑃 𝑔 𝑋1 , … , 𝑋𝑝 ≤ 𝑦 , −∞ < 𝑦 < ∞,
∞ ∞
1 1 𝑛 2
𝐹𝑌 𝑦 = ⋯ 𝑛 𝑒 −2 𝑖=1 𝑥 𝑖 𝑑𝑥1 𝑑𝑥2 ⋯ 𝑑𝑥𝑛 .
−∞ −∞ 2𝜋 2
𝑥 12 +⋯ +𝑥 𝑛2 ≤𝑦
Clearly evaluating the above integral may be tedious. This points towards desirability, if
possible, of other methods of determining the distributions of functions of random
variable. We will see that other techniques are available and, in a given situation, often
one technique is more elegant than the others.