Mathematics Asked on January 5, 2022
Let $(X_1, Y_1)$, and $(X_2, Y_2)$ be two pairs of random variables, and they are assumed to be independent.
Does it mean that:
Motivation:
I had this question because in a lot of applications in statistics, you would like to find a parameter such that the joint distribution
$$P_{Y_1, Y_2}(y_1, y_2| x_1, x_2; theta)$$ is maximized. And there is often the assumption that $(X_1, Y_1)$, and $(X_2, Y_2)$ are independent and identically distributed.
Using this assumption ,
$$P_{Y_1, Y_2}(y_1, y_2| x_1, x_2; theta) = P_{Y_1}(y_1| x_1; theta)P_{Y_1}(y_2| x_2; theta) = prod_{i = 1}^2 P_{Y_i}(y_i| x_i; theta)$$
However, for this calculation to work, you must be able to show,
$$P_{Y_1, Y_2}(y_1, y_2| x_1, x_2; theta) = dfrac{Pr(Y_1 = y_1, Y_2 = y_2, X_1 = x_1, X_2 = x_2; theta)}{Pr(X_1 = x_1, X_2 = x_2; theta)}$$
which means you must split the denominator $Pr(X_1 = x_1, X_2 = x_2; theta) = Pr(X_1 = x_1; theta)Pr(X_2 = x_2; theta)$. However, the independence of $X_1, X_2$ is not explicitly stated. So it is implicit in the assumption? Or is an assumption missing?
I agree that the answers to your four questions are all "YES" (provided the typo in question 3 is fixed).
However, in view of your motivation, I'd like to point out that the way to derive the factorisation equation is like this. First, in view of the independence between $(x_1, y_1)$ and $(x_2, y_2)$, we have $$ p(x_1, x_2, y_1, y_2) = p(x_1, y_1) p(x_2, y_2).$$
Next, it's generally the case that $p(x_1, x_2, y_1, y_2) = p(y_1, y_2 | x_1, x_2) p(x_1, x_2)$ and $p(x_1, y_1) = p(y_1 | x_1)p(x_1)$ and $p(x_2, y_2) = p(y_2 | x_2) p(x_2)$. So $$ p(y_1, y_2 | x_1, x_2) p(x_1, x_2) = p(y_1 | x_1)p(x_1) p(y_2 | x_2) p(x_2)$$
Using independence between $x_1$ and $x_2$ once more (this is like what you were asking for in your denominator), we have $p(x_1, x_2) = p(x_1)p(x_2)$, so $$ p(y_1, y_2 | x_1, x_2) p(x_1) p(x_2) = p(y_1 | x_1)p(x_1) p(y_2 | x_2) p(x_2).$$
Cancelling factors of $p(x_1)p(x_2)$ gives $$ p(y_1, y_2 | x_1, x_2) = p(y_1 | x_1) p(y_2 | x_2) ,$$ which is what you were after.
Edit: To address the comment below...
Saying that $(x_1, y_1)$ and $(x_2, y_2)$ are independent is to say that $$ p(x_1, x_2, y_1, y_2) = p(x_1, y_1) p(x_2, y_2).$$
Integrating both sides with respect to $y_1$ and $y_2$, we have $$ p(x_1, x_2) = iint dy_1 dy_2 p(x_1, x_2, y_1, y_2) = iint dy_1 dy_2 p(x_1, y_1) p(x_2, y_2) \ = int dy_1 p(x_1, y_1) times int dy_2 p(x_2, y_2) = p(x_1) p(x_2), $$ which is what you wanted to prove.
Answered by Kenny Wong on January 5, 2022
YES, YES, NO and YES. If two sigma algebras are independent then any sub-sigma algebras of these are also independent. This proves 1) 2) and 4).
For 3) take two independent (non-constant) random variables $U$ and $V$ and take $X_1=Y_1=U, X_2=Y_2=V$.
Answered by Kavi Rama Murthy on January 5, 2022
Get help from others!
Recent Answers
Recent Questions
© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP