Documente Academic
Documente Profesional
Documente Cultură
I.
INTRODUCTION
wj =
1 d
k
(1)
j =1
y r =
(2)
y r y r .
yr
(3)
j =1
er =
n
1
er .
n r =1
(4)
w (t ) =
At
1,
e
(5)
c (t )
(6)
c (t )
t =1
C. Instance-relevant Combination
Unknown instance is estimated as long as all the base kNN predictors have been established. A regular way refers
the mean or median of the individual predicted values to
composite result. Some other improved methods put forward
weighted combination by consulting the performance of each
base predictor during the training procedure. The one
producing slighter error receives heavier weight and plays a
significant role in combination, whatever the coming
unknown instance is like. Although these strategies cost less,
they can hardly benefit from the diverse characteristics
belonging to various unknown instances.
To make full use of the characteristics, we present an
instance-relevant weighted combination based on the
assumption that the unknown instance should obtain reliable
prediction if its neighbors found in training set have been
estimated accurately by the same base k-NN predictor. In
8
9
10
11
12
13
14
15
16
17
18
19
20
A. Settings
The three public datasets [12] consist of numeric and
categorical attributes, involving different fields about
automobile, heart disease and cloud seeding. The size of
each dataset is given in Table I. For example, auto93 dataset
comprises 82 instances and each instance has a total of 23
attributes, including 6 categorical ones, 16 numeric ones and
one target attribute. We remove a few instances because of
missing values. In fact, k-NN learning is capable of
incomplete instances by ignoring the attributes in distance
calculation. However, this may be unfavorable to attribute
selection. A traditional k-NN prediction algorithm is
implemented as well to be a basis for evaluation. Its only
parameter k is determined by leave-one-out learning in
training procedure.
Dt Dt ( U ), D D( U );
use the datasets projected on U
Predict the instances in Dt , searching the k neighbors from D .
Calculate the mean error (u);
end for
m=min((u)), u1|U|
if m p then
UUmin; Umin is the attribute set corresponding to m
pm, go to 6; continue the selection
else
terminate the attribute selection. // the error can not be reduced
anymore
end if
err(k)p;
end for
, k1Kmax; construct ht with Kt and its
K t = arg min (err (k ))
k
attribute subset At
21 end for; // T base predictors have been built
22 for i=1T // begin to predict (x, y)
23 calculate the predicted value y(i) of (x, y) on hi;
24 calculate the weight w(i) of hi with (5) and (6);
25 end for
T
26
y = w(i ) y (i )
TABLE I.
Dataset
i =1
Attributes
categorical
numeric
total
auto93
82
16
23
Cleveland
297
14
cloud
108
Instances
B. Results
The experimental results on auto93 dataset are shown in
Table II. We use I-R, Mean, Med to indicate different kinds
of combination rules in bagging-based k-NN prediction, i.e.,
instance-relevant, mean and median, whereas k-NN refers to
the traditional algorithm.
Compared with standard k-NN prediction, the mean
relative error has been reduced effectively by the ensemble.
In each experiment round, the error decreases gradually
accompanied with increasing value of T. This improvement
benefits from more candidate base k-NN predictors when
greater T is preset. In this case an unknown instance stands a
good chance of finding out appropriate base predictors with
uniform characteristics. Among the three rules, instancerelevant (I-R) combination surpasses the other two
substantially as the weight of each base predictor is assigned
dynamically in terms of the unknown instance. Although the
performance of both mean and median combination is
inferior to I-R, they do promote k-NN prediction. In
particular, the mean relative error has been reduced by
16.47%-25.25%, according to instance-relevant combination
when T=15, achieving the best results on auto93 dataset.
EXPERIMENTAL RESULTS
In this section experiments are made to test the baggingbased k-NN numeric prediction algorithm. Firstly we
illustrate how the performance is evaluated. Then, the
experimental results are investigated.
others may produce very large errors. Since the ensemble kNN algorithm aims to provide accurate prediction to all
instances, not just part of them, the mean error as well as
standard deviation incline to be reduced consequentially.
Round
Error
T=5
I-R
1
2
3
Mean
T=10
Med
I-R
Mean
T=15
Med
I-R
Mean
Med
k-NN
MREa (%)
17.98
18.39
19.35
16.87
18.24
18.62
16.65
18.05
18.18
19.95
Std. Dev. b
0.1352
0.1336
0.1394
0.1348
0.1357
0.1383
0.1300
0.1337
0.1350
0.1619
MRE (%)
18.97
19.10
20.20
17.75
18.74
19.34
17.09
18.72
19.24
20.46
0.1589
Std. Dev.
0.1459
0.1468
0.1561
0.1451
0.1574
0.1535
0.1332
0.1545
0.1505
MRE (%)
17.67
17.85
18.40
15.51
16.81
17.26
15.25
17.39
17.42
20.40
Std. Dev.
0.1291
0.1311
0.1427
0.1264
0.1255
0.1357
0.1145
0.1228
0.1360
0.1399
Round
Error
T=5
I-R
1
2
3
Mean
T=10
Med
I-R
Mean
T=15
Med
I-R
Mean
Med
k-NN
MRE (%)
29.13
33.35
32.30
29.09
33.02
31.26
27.88
33.19
30.79
34.25
Std. Dev.
0.4275
0.4320
0.4970
0.3916
0.4288
0.4837
0.3752
0.4209
0.4831
0.4637
MRE (%)
31.19
35.78
33.93
29.01
35.60
33.67
28.78
35.22
33.47
36.27
Std. Dev.
0.4936
0.4937
0.5544
0.4482
0.4876
0.5426
0.4418
0.4642
0.5263
0.5831
MRE (%)
25.91
33.43
32.06
25.80
32.89
30.54
25.81
32.85
31.31
34.32
Std. Dev.
0.3922
0.3931
0.4046
0.3936
0.3799
0.4024
0.3922
0.3692
0.4353
0.4919
Round
1
2
3
Error
T=5
T=10
T=15
k-NN
I-R
Mean
Med
I-R
Mean
Med
I-R
Mean
Med
MRE (%)
12.30
12.47
12.67
11.89
12.12
12.14
11.79
12.22
12.29
20.55
0.1453
Std. Dev.
0.1028
0.1074
0.1087
0.0991
0.1026
0.1051
0.0986
0.1036
0.1068
MRE (%)
11.89
12.28
12.43
11.97
12.29
12.20
11.84
12.09
12.37
21.53
Std. Dev.
0.1062
0.1085
0.1124
0.1074
0.1095
0.1119
0.1078
0.1080
0.1126
0.1652
MRE (%)
12.68
12.78
13.11
12.65
12.62
12.73
12.34
12.57
12.50
22.90
Std. Dev.
0.1068
0.1042
0.1065
0.1058
0.1041
0.1062
0.1058
0.1033
0.1071
0.1629
Dataset
Error
T=5
I-R
auto93
Cleveland
cloud
IV.
Mean
T=10
Med
I-R
Mean
T=15
Med
I-R
Mean
Med
k-NN
MRE (%)
18.21
18.45
19.32
16.71
17.93
18.41
16.33
18.05
18.28
20.27
Std. Dev.
0.1367
0.1372
0.1461
0.1354
0.1395
0.1425
0.1259
0.1370
0.1405
0.1536
MRE (%)
28.74
34.19
32.76
27.97
33.84
31.82
27.49
33.75
31.86
34.95
Std. Dev.
0.4378
0.4396
0.4853
0.4111
0.4321
0.4762
0.4031
0.4181
0.4816
0.5129
MRE (%)
12.29
12.51
12.74
12.17
12.34
12.36
11.99
12.29
12.39
21.66
Std. Dev.
0.1053
0.1067
0.1092
0.1041
0.1054
0.1077
0.1041
0.1020
0.1089
0.1578
[1]
CONCLUSIONS