ISI MStat PSB 2004 Problem 7 | Finding the Distribution of a Random Variable

This is a very beautiful sample problem from ISI MStat PSB 2004 Problem 7 based on finding the distribution of a random variable . Let's give it a try !!

Problem- ISI MStat PSB 2004 Problem 7


Suppose X has a normal distribution with mean 0 and variance 25 . Let Y be an independent random variable taking values -1 and 1 with equal probability. Define S=XY+XY and T=XYXY
(a) Find the probability distribution of s.
(b) Find the probability distribution of (S+T10)2

Prerequisites


Cumulative Distribution Function

Normal distribution

Solution :

(a) We can write \( S = \begin{cases} 2x & , if Y=1 \\ -2x & , if Y=-1 \end{cases} \)

Let Cumulative distribution function of S be denoted by \( F_{S}(s) \) . Then ,

\( F_{S}(s) = P(S \le s) = P(S \le s | Y=1)P(Y=1) + P(S \le s| Y=-1)P(Y=-1) = P(2X \le s) \times \frac{1}{2} + P(-2X \le s) \times \frac{1}{2} \) ----(1)

Here given that Y takes values 1 and -1 with equal probabilities .so , \( P(Y=1)=P(Y=-1)= \frac{1}{2} \) .

Now as \(X \sim N(0, 5^2)\) hence X is symmetric distribution about 0 . Thus X and -X are identically distributed .

Thus from (1) we get \( F_{S}(s) = P(X \le s/2 ) \times \frac{1}{2} + P(-X \le s/2) \times \frac{1}{2} = P(X \le s/2 ) \times \frac{1}{2} + P(X \le s/2) \times \frac{1}{2}\)=\( P( X \le s/2) = P(\frac{X-0}{5} \le s/2 ) = \Phi(\frac{s-0}{10}) \)

Hence \( S \sim N(0,{10}^2) \).

(b) Let K=\( (\frac{S+T}{10})^{2}\) = \( \frac{{XY}^2}{ {(10)}^2} \)

Let C.D.F of K be \( F_{K}(k) = P(K \le k ) = P(K \le k | Y=1)P(Y=1) + P(K \le k| Y=-1)P(Y=-1) = P( \frac{X^2}{{(10)}^2} \le k ) \)

=\( P( -10 \sqrt{k} \le X \le 10 \sqrt{k} ) = \Phi(2\sqrt{k}) - \Phi(-2 \sqrt{k}) \) as \(X \sim N(0, 5^2)\).


Food For Thought

Find the distribution of T .


Similar Problems and Solutions



ISI MStat PSB 2008 Problem 10
Outstanding Statistics Program with Applications

Outstanding Statistics Program with Applications

Subscribe to Cheenta at Youtube


Useless Data, Conditional Probability, and Independence | Cheenta Probability Series

This concept of independence, conditional probability and information contained always fascinated me. I have thus shared some thoughts upon this.

When do you think some data is useless?

Some data/ information is useless if it has no role in understanding the hypothesis we are interested in.

We are interested in understanding the following problem.

\(X\) is some event. \(Y\) is another event. How much information do \(Y\) and \(X\) give about each other?

We can model an event by a random variable. So, let's reframe the problem as follows.

\(X\) and \(Y\) are two random variables. How much information do \(Y\) and \(X\) give about each other?

There is something called entropy. But, I will not go into that. Rather I will give a probabilistic view only. The conditional probability marches in here. We have to use the idea that we have used the information of \(Y\), i.e. conditioned on \(Y\). Hence, we will see how \(X \mid Y\) will behave?

How does \( X \mid Y\) behave? If \(Y\) has any effect on \(X\), then \(X \mid Y\) would have changed right?

But, if \(Y\) has no effect on \(X\), then \(X \mid Y\) will not change and remain same as X. Mathematically, it means

\( X \mid Y\) ~ \(X\) \(\iff\) \( X \perp \!\!\! \perp Y\)

We cannot distinguish between the initial and the final even after conditioning on \(Y\).

Theorem

\(X\) and \(Y\) are independent \( \iff \) \( f(x,y) = P(X =x \mid Y = y) \) is only a function of \(x\).

Proof

\( \Rightarrow\)

\(X\) and \(Y\) are independent \( \Rightarrow \) \( f(x,y) = P(X =x \mid Y = y) = P(X = x)\) is only a function of \(x\).

\( \Leftarrow \)

Let \( \Omega \) be the support of \(Y\).

\( P(X =x \mid Y = y) = g(x) \Rightarrow \)

\( P(X=x) = \int_{\Omega} P(X =x \mid Y = y).P(Y = y)dy \)

\(= g(x) \int_{\Omega} P(Y = y)dy = g(x) = P(X =x \mid Y = y) \)

Exercises

  1. \((X,Y)\) is a bivariate standard normal with \( \rho = 0.5\) then \( 2X - Y \perp \!\!\! \perp Y\).
  2. \(X, Y, V, W\) are independent standard normal, then \( \frac{VX + WY}{\sqrt{V^2+W^2}} \perp \!\!\! \perp (V,W) \).

Random Thoughts (?)

How to quantify the amount of information contained by a random variable in another random variable?

Information contained in \(X\) = Entropy of a random variable \(H(X)\) is defined by \( H(X) = E(-log(P(X)) \).

Now define the information of \(Y\) contained in \(X\) as \(\mid H(X) - H(X|Y) \mid\).

Thus, it turns out that \(H(X) - H(X|Y) = E_{(X,Y)} (log(\frac{P(X \mid Y)}{P(X)})) = H(Y) - H(Y|X) = D(X,Y)\).

\(D(X,Y)\) = Amount of information contained in \(X\) and \(Y\) about each other.

Exercise

Note: This is just a mental construction I did, and I am not sure of the existence of the measure of this information contained in literature. But, I hope I have been able to share some statistical wisdom with you. But I believe this is a natural construction, given the properties are satisfied. It will be helpful, if you get hold of some existing literature and share it to me in the comments.

Some useful Links: