Jethro's Braindump

Generalisation Error

We can define generalisation error as the discrepancy between Ein and Eout. The Hoeffding Inequality characterises the generalization error with a probabilistic bound:

P[|Ein(g)Eout(g)|>ϵ]2Me2ϵ2N

Pick a tolerance level δ, and assert with probability 1δ that

Eout(g)Ein(g)+12Nln2Mδ

Notice the error bound depends on M, the size of the hypothesis set H. Most learning models have infinite H, including the simple perceptron. Hence, to study generalisation in such models, we need to derive a counterpart that deals with infinite H.

Notice that the M factor was obtained by taking the disjunction of events. Let Bm be the bad event that |Ein(hm)Eout(hm)|>ϵ. Notice that these bad events are often strongly overlapping, and the disjunction of these events form a much smaller area.

The mathematical theory of generalisation hinges on this observation. Upon accounting for the overlaps of different hypotheses, we will be able to replace the number of hypotheses M with an effective finite number, even while M is infinite.

Links to this note