menu

OPT_2: Gradient descent convergence rate (P1)

date_range 21/12/2021 15:50

Gradient descend convergence guarantee

  • Given an objective function with is the parameters (weights), satisfies:
    • Lipschitz continuous:
    • function:

UML_6: Learning via Uniform Convergence

date_range 24/11/2021 09:36

UML I.4 Learning via Uniform Convergence

  • Recall in previous posts, we discussed about the realizable assumption and ERM learning. We hope an hypothesis , when minimizing error on , also respect to . In other words, we need all members of are good approximations of their true risk.
  • Def 1(-representative sample): A training set is called -representative if
  • Lemma 1: Assume that a training set is -representative, any satisfies
    • This lemma implies that the ERM rule is an agnostic PAC learner, it suffices to show that with probability of at least over the random choice of a , it will be an -representative training set.
    • The proof in .

OPT_1: Gradient descent on convex function

date_range 02/07/2021 03:50

  • As we know in gradient descent method, we shift our parameters against descent direction. In this post, we denote as:

MATH_7: From Binomial to Poisson distribution

date_range 22/04/2021 02:59

math7

1. Binomial distribution
  • Problem: Toss a coin times, let represents "obtain a head" and represents "obtain a tail" where . The probability of obtaining times is:

CVX_3: Operations that preserve convexity

date_range 15/04/2021 19:20

cvx3 In this post, we study about the operations on sets that preserve convexity.

1. Intersection
  • Theorem 1: If all of are convex, is convex. ()