Here we can see how the k influence the classification itself. We can see that with increasing k the
classification rate is decreasing (figure 4.7 on page ) and that there exist some k for which the
classification is the best. This is not necessarily the minimum one as we can see that for SIFTs with worse
discrimination power (number of square patches: 6x3, 6x4) the peak is for the
.
For SIFTs with number of square patches 8x4 and 15x9 the peak is for
and
respectively.
|
Let us show what is happening on the picture figure 4.8 on page . Let
is a test car and
are
car classes from the learning set. The classification error for
is
, for
is
.
Leat us assume there is an
which denotes a threshold for which the assignment between
is pointless. In another words
has nothing to do with
.
If we choose some
we are trying to find a class which has a majority between
observations. The problem is that if from some
the error
exists we are taking
into account car class
which have noting to do with
.
The graph on figure 4.7 on page is showing the classification is going worse from
.
The problem was if we had e.g. 1x Fiat Uno with 2 total occurences in class and 4x Skoda Felicie with total
number of occurences
. Therefore the relative majority is for Fiat Uno since it has relative occurence
of 0.5 whereas Felicie would have at most
.
Kocurek 2007-12-17