UML_4: PAC Learning | 3.1.
-
date_range 26/04/2020 23:23 infosortStatistical_Learninglabelslvapnik
3.1. PAC learning
In UML_3, we have shown that for a finite
hypothesis class and a sufficiently large training sample (follow the distribution and labeling function ) then output of will be probably approximately correct. Now, we define Probably Approximately Correct (PAC) learning. Define (PAC learning):
is PAC learnable if there exist a function with property: - For every
, every over and every , the relizable assumption (Giả thuyết tính khả thi in UML_3) holds with respect , then when running the learning algorithm on , the algorithm returns a hypothesis such that, with probability of at least that we have .
- For every
In PAC learning, we have two parameters:
determines how far the output classifier can be optimal. indicating how likely meets accuracy .
Sample complexity:
The function
determines the sample complexity of . How many samples are required to guarantee property of PAC? It also depend on the size of . If
is PAC learnable, there are many that satisfy the requirement given in the Define. Thus, we will define the sample complexity of so that for any , can satisfy the requirement of PAC. Let remember the conclusion of the analysis of finite hypothesis class in UML_3. Now we have: Corollary: Every finite hypothesis
is PAC learnable with sample complexity: