Documente Academic
Documente Profesional
Documente Cultură
Luis De Alba
ldealbar@cc.hut.fi
5.3.2008
Boosting
Definition:
Method of producing a very accurate
prediction rule by combining
inaccurate rules of thumb.
Objective:
Improve the accuracy of any given
learning algorithm (LMS, SVM,
Perceptron, etc.)
How to boost in 2 steps:
Select a Classifier with an accuracy
greater than chance (random)
Form an ensemble of the selected
Classifiers whose joint decision rule
has high accuracy.
Example 1: Golf Tournament
Tiger Woods VS
18 average+ players
Example 2: Classification Problem
2 dimensions, 2 classes, 3 Classifiers
D training sets {x1,y1 ... xm,ym}
a) Train C1 with D1 where D1 is
randomly selected.
b) Train C2 with D2 where only 50%
of the selected set is correctly
classified by C1
c) Train C3 with D3 where the
classification, of selected D3,
disagrees between C1 and C2
Classification task:
C1
Y, C1
X C2 C1=C2 y
N, C3
C3
Duda, et al.
AdaBoost
x1 y1 x2 y2 x3 y3 x4 y4 x5 y5 x6 y6
Ck
Increase weights of training
patterns misclassified by C k
Decrease weights of training
patterns correctly classified by C k
w2 w4
w1 w3 w5 w6
x1 y1 x2 y2 x3 y3 x4 y4 x5 y5 x6 y6
W k1 i
Zk
return
C k and k for all K
Output of the final classifier for any
given input.
K
H x =sign k C k x
k=1
Analysis
m
Error calculation: if C k x i y i : 1 : 0
E k= i=1
m
Error is the summation and
normalization of all wrongly classified
sets by the weak learner.
Weak learner shall do better than if
only random guesses i.e., E k0.5
measurement:
1
k ln
2 [
1E k
Ek ]
It measures the importance
assigned to C k
There are two things to note:
>=0 if E <=
gets larger as E gets smaller
Normalization factor:
Z k is a normalization factor so that W k1
will be a distribution.
Possible calculation:
m
Z k = W k1 i
i=1
Hypothesis:
K
H x =sign k Ck x
k =1
y2
x1 x2
y3
y4
y5
Decompose the problem into n
orthogonal binary classification
problems.
For Y , define Y [] for to be
{
Y []= 1 if Y
1 if Y
H : XY {1,1} defined by H x ,=H x []
Then, replace each training
example by k examples.
x i ,Y i x i , ,Y i [] for