menu

UML_4: PAC Learning | 3.1.

3.1. PAC learning

3.1. PAC learning
  • In UML_3, we have shown that for a finite hypothesis class and a sufficiently large training sample (follow the distribution and labeling function ) then output of will be probably approximately correct. Now, we define Probably Approximately Correct (PAC) learning.

  • Define (PAC learning): is PAC learnable if there exist a function with property:

    • For every , every over and every , the relizable assumption (Giả thuyết tính khả thi in UML_3) holds with respect , then when running the learning algorithm on , the algorithm returns a hypothesis such that, with probability of at least that we have .
  • In PAC learning, we have two parameters:

    • determines how far the output classifier can be optimal.
    • indicating how likely meets accuracy .
  • Sample complexity:

    • The function determines the sample complexity of . How many samples are required to guarantee property of PAC? It also depend on the size of .

    • If is PAC learnable, there are many that satisfy the requirement given in the Define. Thus, we will define the sample complexity of so that for any , can satisfy the requirement of PAC. Let remember the conclusion of the analysis of finite hypothesis class in UML_3. Now we have:

      Corollary: Every finite hypothesis is PAC learnable with sample complexity:

Reference: