CVX_2: Convex set and Cone
cvx2
date_range 07/04/2021 00:20
1. Convex sets
- Definition 1: A set is convex if the line segment between any two distinct points in lies in .
- For any , , , we have .
CVX_1: Affine set
cvx1
date_range 15/03/2021 16:03
1. Line and line segment:
- Suppose we have two points :
- The line through two this points:
- The line segment between two this points:
- The line through two this points:
ML&PR_10: Linear Regression in statistic view
mlpr1x
date_range 14/01/2021 16:03
- In this post, we discuss about Linear Regression in statistic view, thereby answer why we should use Mean Square Error as a loss function.
- We suppose that target variable is given by a deterministic function with additive Gaussian noise :
where is a zero mean Gaussian random variable with precision . We can write:
- Now consider a data set of inputs with corresponding target values . We have:
- We wish to maximize , synonymous with maximize :
wherewe must minimize - the MSE loss. And we are done!
ML&PR_9: Bayes’ theorem for Gaussian variables
mlpr9
date_range 05/01/2021 16:02
- In this post, we discuss about Bayes' theorem for Gaussian variables. With
given and a Gaussian conditional distribution , we wish to find . As show in MLPR8, has mean is a linear function of :
- We consider the joint distribution:
Take the log, we obtain:
- In order to find covariance matrix of , we rewrite the as a quadratic function:
- Finally, we get:
ML&PR_8: Conditional Gaussian
mlpr8
date_range 30/11/2020 23:02
- Consider two sets of variables are jointly Gaussian, then, the conditional distribution of one set conditioned on the other is again Gaussian.
- Suppose is a -dimensional vector with Gaussian distribution . We split into two parts: and where takes first components of and takes remaining components.
- We now define the mean of :
and the covariance matrix of :In there, , and are symmetric and .
- We now define precision matrix takes a form:
Because is symmetric, also is symmetric . So we can rewrite this matrix as follows:where , and are symmetric and . Note that: is not the invert of , similar to . We will discuss about it later.
- Now we discuss about the conditional distribution , consider is the observed value. We start form the joint distribution . To explore it, we consider the quadratic form of Gaussian distribution (as mentioned in MLPR8) combine with the partitioning and :
The is the function of , we can use this property(because of is symmetric then ) to rewrite it by:whereWe obtain from by property . We see this is again a quadratic form with is independent of . So the condition distribution will be Gaussian.
- We use the quadratic form to determine mean and covariance of conditional distribution, denoted as and , respectively. Consider the second order of :
we can inference the covariance of is:Next, we consider the first order of :we can obtain:
- Next, we determine each path of based on the following lemma:
where is the Schur complement of and is the Schur complement of , defined as:
- Back to we have:
So: