ML&PR_8: Conditional Gaussian
-
date_range 30/11/2020 23:02 infosortMachine_Learning_n_Pattern_Recognitionlabelmlbishopstat
Consider two sets of variables are jointly Gaussian, then, the conditional distribution of one set conditioned on the other is again Gaussian.
Suppose is a -dimensional vector with Gaussian distribution . We split into two parts: and where takes first components of and takes remaining components.
We now define the mean of :
and the covariance matrix of :
In there, , and are symmetric and .
We now define precision matrix takes a form:
Because is symmetric, also is symmetric . So we can rewrite this matrix as follows:
where , and are symmetric and .
Note that: is not the invert of , similar to . We will discuss about it later.
Now we discuss about the conditional distribution , consider is the observed value. We start form the joint distribution . To explore it, we consider the quadratic form of Gaussian distribution (as mentioned in MLPR8) combine with the partitioning and :
The is the function of , we can use this property
(because of is symmetric then )
to rewrite it by:
where
We obtain from by property . We see this is again a quadratic form with is independent of . So the condition distribution will be Gaussian.
We use the quadratic form to determine mean and covariance of conditional distribution, denoted as and , respectively. Consider the second order of :
we can inference the covariance of is:
Next, we consider the first order of :
we can obtain:
Next, we determine each path of based on the following lemma:
where is the Schur complement of and is the Schur complement of , defined as:
Back to we have:
So:
Appendix:
The inverse of a symmetric matrix also symmetric.
Proof:
Given the inversible and symmetry matrix :
, we will prove:
First, we have:
From , we obtain:
Then, use two properties and :
Finally, we put into right side of both:
and we are done.
In order to prove formula , we consider the follow equation:
It equivalent to:
From the under equation of the formula , we obtain:
Replace the into the , we have:
where .
Replace the into the , we calculate the :
From the and the , we have:
As mentioned at , we also have:
Then, we achieve the goal.
Reference:
- 2.3.1| Pattern Recognition and Machine Learning | C.M. Bishop.
- The Schur Complement and Symmetric PositiveSemidefinite (and Definite) Matrices.