m (Protected "ECE600 F13 Joint Distributions mhossain" [edit=sysop:move=sysop])
 
(10 intermediate revisions by 2 users not shown)
Line 3: Line 3:
 
[[ECE600_F13_notes_mhossain|Back to all ECE 600 notes]]
 
[[ECE600_F13_notes_mhossain|Back to all ECE 600 notes]]
  
 +
 +
[[Category:ECE600]]
 +
[[Category:probability]]
 +
[[Category:lecture notes]]
 +
[[Category:slecture]]
  
 
<center><font size= 4>
 
<center><font size= 4>
'''Random Variables and Signals'''
+
[[ECE600_F13_notes_mhossain|'''The Comer Lectures on Random Variables and Signals''']]
 
</font size>
 
</font size>
 +
 +
[https://www.projectrhea.org/learning/slectures.php Slectures] by [[user:Mhossain | Maliha Hossain]]
 +
  
 
<font size= 3> Topic 11: Two Random Variables: Joint Distribution</font size>
 
<font size= 3> Topic 11: Two Random Variables: Joint Distribution</font size>
 
</center>
 
</center>
  
 
+
----
 
----
 
----
  
Line 29: Line 37:
  
  
but this would not capture the joint behavior or X and Y. Note also that if X and Y are defined on two different probability spaces, those two spaces can be combined to create (''S,F'',P).
+
but this would not capture the joint behavior of X and Y. Note also that if X and Y are defined on two different probability spaces, those two spaces can be combined to create (''S,F'',P).
  
 
In order for X and Y to be a valid random variable pair, we will need to consider regions D ⊂ '''R'''<math>^2</math>.<br/>
 
In order for X and Y to be a valid random variable pair, we will need to consider regions D ⊂ '''R'''<math>^2</math>.<br/>
Line 35: Line 43:
  
 
We need {(X,Y) ∈ O} ∈ ''F'' for any open rectangle O ⊂ '''R'''<math>^2</math>, then {(X,Y) ∈ D} ∈ ''F'' ∀D ∈ B('''R'''<math>^2</math>).<br/>
 
We need {(X,Y) ∈ O} ∈ ''F'' for any open rectangle O ⊂ '''R'''<math>^2</math>, then {(X,Y) ∈ D} ∈ ''F'' ∀D ∈ B('''R'''<math>^2</math>).<br/>
But (X(<math>\omega</math>),Y(<math>\omega</math>)) ∈ O if X(<math>\omega</math> ∈ A and Y(<math>\omega</math> ∈ B for some A, B ∈ B('''R'''), so {(X,Y) ∈ 0} = X<math>^{-1}</math>(A) ∩ Y<math>^{-1}</math>(B)<br/>
+
But (X(<math>\omega</math>),Y(<math>\omega</math>)) ∈ O if X(<math>\omega</math>) ∈ A and Y(<math>\omega</math>) ∈ B for some A, B ∈ B('''R'''), so {(X,Y) ∈ O} = X<math>^{-1}</math>(A) ∩ Y<math>^{-1}</math>(B)<br/>
 
If X and Y are valid random variables then <br/>
 
If X and Y are valid random variables then <br/>
 
<center><math>\begin{align}
 
<center><math>\begin{align}
Line 59: Line 67:
 
Knowledge of F<math>_X</math>(x) and F<math>_Y</math>(y) alone will not be sufficient to compute P((X,Y) ∈ D) ∀D ∈ B('''R'''<math>^2</math>), in general.  
 
Knowledge of F<math>_X</math>(x) and F<math>_Y</math>(y) alone will not be sufficient to compute P((X,Y) ∈ D) ∀D ∈ B('''R'''<math>^2</math>), in general.  
  
'''Definition''' <math>\qquad</math> The '''joint cumulative distribution function''' of random variables X,Y defined on (''S,F'',P) is F<math>_{XY}</math>(x,y) ≡ P({X ≤ x}{Y ≤ y}) for x,y ∈ '''R'''.<br/>
+
'''Definition''' <math>\qquad</math> The '''joint cumulative distribution function''' of random variables X,Y defined on (''S,F'',P) is F<math>_{XY}</math>(x,y) ≡ P({X ≤ x} {Y ≤ y}) for x,y ∈ '''R'''.<br/>
 
Note that in this case, D ≡ D<math>_{XY}</math> = {(x',y') ∈ '''R'''<math>^2</math>: x' ≤ x, y' ≤ y}
 
Note that in this case, D ≡ D<math>_{XY}</math> = {(x',y') ∈ '''R'''<math>^2</math>: x' ≤ x, y' ≤ y}
  
Line 80: Line 88:
  
 
'''Definition''' <math>\qquad</math> The '''joint probability density function''' of random variables X and Y is <br/>
 
'''Definition''' <math>\qquad</math> The '''joint probability density function''' of random variables X and Y is <br/>
<center><math>f_{XY}(x,y) \equiv \frac{\delta^2}{\delta x\delta y}F_{XY}(x,y)</math></center>
+
<center><math>f_{XY}(x,y) \equiv \frac{\partial^2}{\partial x\partial y}F_{XY}(x,y)</math></center>
  
 
∀(x,y) ∈ '''R'''<math>^2</math> where the derivative exists.
 
∀(x,y) ∈ '''R'''<math>^2</math> where the derivative exists.
Line 86: Line 94:
 
It can be shown that if D ∈ B('''R'''<math>^2</math>), then, <br/>
 
It can be shown that if D ∈ B('''R'''<math>^2</math>), then, <br/>
 
<center><math>P((X,Y)\in D)=\int\int_Df_{XY}(x,y)dxdy</math></center>
 
<center><math>P((X,Y)\in D)=\int\int_Df_{XY}(x,y)dxdy</math></center>
 +
 +
where D ≡ D<math>_{XY}</math> = {(x',y') ∈ '''R'''<math>^2</math>: x' ≤ x, y' ≤ y}
  
  
 
==Properties of f<math>_{XY}</math>:==
 
==Properties of f<math>_{XY}</math>:==
 
<math>\bullet f_{XY}(x,y)\geq 0\qquad\forall x,y\in\mathbb R</math><br/>
 
<math>\bullet f_{XY}(x,y)\geq 0\qquad\forall x,y\in\mathbb R</math><br/>
<math>\bullet \int\int_{\mathbb R}f_{XY}(x,y)dxdy</math>
+
<math>\bullet \int\int_{\mathbb R}f_{XY}(x,y)dxdy = 1</math><br/>
<math>\bullet F_{XY}(x,y) = \int_{-\infty}^{y}\int_{-\infty}^xf_{XY}(x',y')dx'dy'\qquad\forall(x,y)\in\mathbb R</math><br/>
+
<math>\bullet F_{XY}(x,y) = \int_{-\infty}^{y}\int_{-\infty}^xf_{XY}(x',y')dx'dy'\qquad\forall(x,y)\in\mathbb R^2</math><br/>
 
<math>\begin{align}
 
<math>\begin{align}
 
\bullet &f_X(x) = \int_{-\infty}^{\infty}f_{XY}(x,y)dy \\
 
\bullet &f_X(x) = \int_{-\infty}^{\infty}f_{XY}(x,y)dy \\
Line 103: Line 113:
  
 
If X and Y are discrete random variables, we will use the joint pdf given by <br/>
 
If X and Y are discrete random variables, we will use the joint pdf given by <br/>
<center><math>p_{XY}(x,y) = P(X=x,Y=y)\qquad \forall(x,y)\in\mathcal R_X \times\mathcal R_X</math></center>
+
<center><math>p_{XY}(x,y) = P(X=x,Y=y)\qquad \forall(x,y)\in\mathcal R_X \times\mathcal R_Y</math></center>
  
Note that if X is discrete and Y continuous (or vice versa), we will be interested in <br>
+
Note that if X is continuous and Y discrete (or vice versa), we will be interested in <br>
<center><math>P(\{X\in A\}\cap\{Y=y\}),\;\;A\in B(\mathbb R);y\in\mathcal R_Y</math></center>
+
<center><math>P(\{X\in A\}\cap\{Y=y\}),\;\;A\in B(\mathbb R),\;y\in\mathcal R_y</math></center>
 
We often use a form of Bayes' Theorem, which we will discuss later, to get this probability.  
 
We often use a form of Bayes' Theorem, which we will discuss later, to get this probability.  
  
Line 123: Line 133:
 
==Special Case==
 
==Special Case==
  
We often model X and Y as jointly Gaussian with μ<math>_X</math> = μ<math>_Y</math> = 0, σ<math>_X</math> = σ<math>_Y</math> = σ, r= = 0, so that <br/>
+
We often model X and Y as jointly Gaussian with μ<math>_X</math> = μ<math>_Y</math> = 0, σ<math>_X</math> = σ<math>_Y</math> = σ, r = 0, so that <br/>
 
<center><math>f_{XY}(x,y) = \frac{1}{2\pi\sigma^2}e^{-\frac{x^2+y^2}{2\sigma^2}}</math></center>
 
<center><math>f_{XY}(x,y) = \frac{1}{2\pi\sigma^2}e^{-\frac{x^2+y^2}{2\sigma^2}}</math></center>
  
  
'''Example''' <math>/qquad</math> Let X and Y be jointly Gaussian with μ<math>_X</math> = μ<math>_Y</math> = 0, σ<math>_X</math> = σ<math>_Y</math> = σ, r= = 0. Find the probability that (X,Y) lies within a distance d from the origin.  
+
'''Example''' <math>\qquad</math> Let X and Y be jointly Gaussian with μ<math>_X</math> = μ<math>_Y</math> = 0, σ<math>_X</math> = σ<math>_Y</math> = σ, r = 0. Find the probability that (X,Y) lies within a distance d from the origin.  
  
 
Let <br>
 
Let <br>
Line 147: Line 157:
 
Then <br/>
 
Then <br/>
 
<center><math>\begin{align}
 
<center><math>\begin{align}
P((X,Y)\in D_d) &= \int_{\pi}^{\pi}\int_{0}^{d}f_{XY}(r\cos\theta,r\sin\theta)rdrd\theta \\
+
P((X,Y)\in D_d) &= \int_{-\pi}^{\pi}\int_{0}^{d}f_{XY}(r\cos\theta,r\sin\theta)rdrd\theta \\
&= \int_{\pi}^{\pi}\int_{0}^{d} \frac{r}{2\pi\sigma^2}e^{-\frac{r^2}{2\sigma^2}}drd\theta \\
+
&= \int_{-\pi}^{\pi}\int_{0}^{d} \frac{r}{2\pi\sigma^2}e^{-\frac{r^2}{2\sigma^2}}drd\theta \\
 
&= 1-e^{-\frac{d^2}{2\sigma^2}}
 
&= 1-e^{-\frac{d^2}{2\sigma^2}}
 
\end{align}</math></center>
 
\end{align}</math></center>

Latest revision as of 12:12, 21 May 2014

Back to all ECE 600 notes

The Comer Lectures on Random Variables and Signals

Slectures by Maliha Hossain


Topic 11: Two Random Variables: Joint Distribution




Two Random Variables

We have been considering a single random variable X and introduces the pdf f$ _X $, and pmf p$ _X $, conditional pdf f$ _X $(x|M), the conditional pmf p$ _X $(x|M), pdf f$ _Y $ or pmf p$ _Y $ when Y = g(X), expectation E[g(X)], conditional expectation E[g(X)|M], and characteristic function $ \Phi_X $. We will now define similar tools for the case of two random variables X and Y.
How do we define two random variables X,Y on a probability space (S,F,P)?


Fig 1: Mapping from S to X($ \omega $) and Y($ \omega $).


So two random variables can be viewed aw a mapping from S to R$ ^2 $, and (X,Y) is an ordered pair in R$ ^2 $. Note that we could draw the picture this way:

Fig 2: Mapping from S to X($ \omega $) and Y($ \omega $). Note that this model does not capture the joint behavior of X and Y and is hence incomplete.


but this would not capture the joint behavior of X and Y. Note also that if X and Y are defined on two different probability spaces, those two spaces can be combined to create (S,F,P).

In order for X and Y to be a valid random variable pair, we will need to consider regions D ⊂ R$ ^2 $.

$ B(\mathbb{R}^2) = \sigma (\{\mbox{all open rectangles in }\mathbb{R}^2\}) $

We need {(X,Y) ∈ O} ∈ F for any open rectangle O ⊂ R$ ^2 $, then {(X,Y) ∈ D} ∈ F ∀D ∈ B(R$ ^2 $).
But (X($ \omega $),Y($ \omega $)) ∈ O if X($ \omega $) ∈ A and Y($ \omega $) ∈ B for some A, B ∈ B(R), so {(X,Y) ∈ O} = X$ ^{-1} $(A) ∩ Y$ ^{-1} $(B)
If X and Y are valid random variables then

$ \begin{align} &X^{-1}(A) \in \mathcal F \\ &Y^{-1}(B) \in \mathcal F \\ &\forall A,B\in B(\mathbb R) \end{align} $

So,

$ \begin{align} &X^{-1}(A)\cap Y^{-1}(B) \in \mathcal F \\ \Rightarrow &\{(X,Y)\in O\}\in\mathcal F \end{align} $

So how do we find P((X,Y) ∈ D) for D ∈ B(R$ ^2 $)?

We will use joint cdfs, pdfs, and pmfs.



Joint Cumulative Distribution Function

Knowledge of F$ _X $(x) and F$ _Y $(y) alone will not be sufficient to compute P((X,Y) ∈ D) ∀D ∈ B(R$ ^2 $), in general.

Definition $ \qquad $ The joint cumulative distribution function of random variables X,Y defined on (S,F,P) is F$ _{XY} $(x,y) ≡ P({X ≤ x} ∩ {Y ≤ y}) for x,y ∈ R.
Note that in this case, D ≡ D$ _{XY} $ = {(x',y') ∈ R$ ^2 $: x' ≤ x, y' ≤ y}

Fig 3: The shaded region represents D


Properties of F$ _{XY} $:


$ \bullet\lim_{x\rightarrow -\infty}F_{XY}(x,y) = \lim_{y\rightarrow -\infty}F_{XY}(x,y) = 0 $
$ \begin{align} \bullet &\lim_{x\rightarrow \infty}F_{XY}(x,y) = F_Y(y)\qquad \forall y\in\mathbb R \\ &\lim_{y\rightarrow \infty}F_{XY}(x,y) = F_X(x)\qquad \forall x\in\mathbb R \end{align} $
F$ _X $ and F$ _Y $ are called the marginal cdfs of X and Y.
$ \bullet P(\{x_1 < X\leq x_2\}\cap\{y_1<Y\leq y_2\}) = F_{XY}(x_2,y_2)-F_{XY}(x_1,y_2)-F_{XY}(x_2,y_1)+F_{XY}(x_1,y_1) $



The Joint Probability Density Function

Definition $ \qquad $ The joint probability density function of random variables X and Y is

$ f_{XY}(x,y) \equiv \frac{\partial^2}{\partial x\partial y}F_{XY}(x,y) $

∀(x,y) ∈ R$ ^2 $ where the derivative exists.

It can be shown that if D ∈ B(R$ ^2 $), then,

$ P((X,Y)\in D)=\int\int_Df_{XY}(x,y)dxdy $

where D ≡ D$ _{XY} $ = {(x',y') ∈ R$ ^2 $: x' ≤ x, y' ≤ y}


Properties of f$ _{XY} $:

$ \bullet f_{XY}(x,y)\geq 0\qquad\forall x,y\in\mathbb R $
$ \bullet \int\int_{\mathbb R}f_{XY}(x,y)dxdy = 1 $
$ \bullet F_{XY}(x,y) = \int_{-\infty}^{y}\int_{-\infty}^xf_{XY}(x',y')dx'dy'\qquad\forall(x,y)\in\mathbb R^2 $
$ \begin{align} \bullet &f_X(x) = \int_{-\infty}^{\infty}f_{XY}(x,y)dy \\ &f_Y(y) = \int_{-\infty}^{\infty}f_{XY}(x,y)dx \end{align} $ are the marginal pdfs of X and Y.



The Joint Probability Mass Function

If X and Y are discrete random variables, we will use the joint pdf given by

$ p_{XY}(x,y) = P(X=x,Y=y)\qquad \forall(x,y)\in\mathcal R_X \times\mathcal R_Y $

Note that if X is continuous and Y discrete (or vice versa), we will be interested in

$ P(\{X\in A\}\cap\{Y=y\}),\;\;A\in B(\mathbb R),\;y\in\mathcal R_y $

We often use a form of Bayes' Theorem, which we will discuss later, to get this probability.



Joint Gaussian Random Variables

An important case of two random variables is: X and Y are jointly Gaussian if their joint pdf is given by

$ f_{XY}(x,y)=\frac{1}{2\pi\sigma_X\sigma_Y\sqrt{1-r^2}}exp\left\{-\frac{1}{2(1-r^2)}\left[\frac{(x-\mu_X)^2}{\sigma_X^2}-\frac{2r(x-\mu_X)(y-\mu_Y)}{\sigma_X\sigma_Y}+\frac{(y-\mu_y)^2}{\sigma_Y^2}\right]\right\} $

where μ$ _X $, μ$ _Y $, σ$ _X $, σ$ _Y $, r ∈ R; σ$ _X $$ _Y $ > 0; -1 <r <1.

It can be shown that is X and Y are jointly Gaussian then X is N(μ$ _X $, σ$ _X $$ ^2 $) and Y is N(μ$ _Y $, σ$ _Y $$ ^2 $) (proof)


Special Case

We often model X and Y as jointly Gaussian with μ$ _X $ = μ$ _Y $ = 0, σ$ _X $ = σ$ _Y $ = σ, r = 0, so that

$ f_{XY}(x,y) = \frac{1}{2\pi\sigma^2}e^{-\frac{x^2+y^2}{2\sigma^2}} $


Example $ \qquad $ Let X and Y be jointly Gaussian with μ$ _X $ = μ$ _Y $ = 0, σ$ _X $ = σ$ _Y $ = σ, r = 0. Find the probability that (X,Y) lies within a distance d from the origin.

Let

$ D_d = \{(x,y)\in\mathbb R^2:\;x^2+y^2\leq d^2\} $


Fig 4:The shaded region shows D$ _d $ = {(x,y)∈R$ ^2 $: x$ ^2 $+y$ ^2 $ ≤ d}


Then

$ P((X,Y)\in D_d) = \int\int_{D_d}\frac{1}{2\pi\sigma^2}e^{-\frac{x^2+y^2}{2\sigma^2}}dxdy $

Use polar coordinates to make integration easier: let

$ \begin{align} r&=x^2+y^2 \\ \theta &= \tan^{-1}(\frac{x}{y}) \end{align} $

Then

$ \begin{align} P((X,Y)\in D_d) &= \int_{-\pi}^{\pi}\int_{0}^{d}f_{XY}(r\cos\theta,r\sin\theta)rdrd\theta \\ &= \int_{-\pi}^{\pi}\int_{0}^{d} \frac{r}{2\pi\sigma^2}e^{-\frac{r^2}{2\sigma^2}}drd\theta \\ &= 1-e^{-\frac{d^2}{2\sigma^2}} \end{align} $


So the probability that (X,Y) lies within distance d from the origin looks like the graph in figure 5 (as a function of d).

Fig 5: P({X$ ^2 $,Y$ ^2 $ ≤ d}) plotted as a function of d



References



Questions and comments

If you have any questions, comments, etc. please post them on this page



Back to all ECE 600 notes

Alumni Liaison

Ph.D. on Applied Mathematics in Aug 2007. Involved on applications of image super-resolution to electron microscopy

Francisco Blanco-Silva