PAC Learning Model
Intro
This model intends to fix the major problem with the consistency model. It should say something about generalizing from a smaller set of data to a larger set of examples.
Generalization Error
Our goal is to obtain an accurate hypothesis. We need to define an error to measure the accuracy.
Prx∼D[h(x)=c(x)]=errD(h)
where x is an example that comes from an unknown target distribution D. h is the hypothesis. testing examples also come from D.
Now we no longer seek a hypothesis that is consistent with training set. Rather, we want to formulate a hypothesis that minimize the errD(h).
Note: Since the training data is randomly selected from an unknown distribution, there is always the chance the training set is very unrepresentative of the source distribution.
The Probably Approximately Correct Model
A target concept class C is PAC-learnable by a hypothesis space H if ∃ an algorithm A such that ∀c∈C, any target distribution D, any positive ϵ and δ, A uses a training set S={(x1,c(x1)),(x2,c(x2),...,(xm,c(xm))} where m=ploy(ϵ1,δ1,...) examples taken from iid from D and produces h∈H such that errD(h)≤ϵ and Pr[errD(h)]≤1−δ
Explanation:
m is a polynomial dependent on ϵ and δ. for a more accurate hypothesis, you need to a larger training set.
ϵ is the accuracy parameter, δ is the confidence parameter. They are both user specified.
We want a hypothesis that is highly probably (1−δ) approximately correct (ϵ-good). The probably approximately correct model, get it?
Last updated