60 6. PROBABILISTIC INTERPRETATION OF CHANNEL REPRESENTATIONS
A second problem are outliers that are located outside the channel representation: two
channel representations might have the same coefficient vectors, but are generated from different
numbers of samples. ese additional samples must influence the posterior distribution.
e first problem requires a de-correlation of the coefficient vector, which is basically the
problem solved in Section 5.3. e calculation of the coefficient vector in (5.22) consist mainly of
solving the linear system with the channel correlation matrix as a system matrix. Consequently,
the obtained coefficient vector
is a good approximation of independent events in the sense of
a histogram. us, we replace c with
in (6.11).
Regarding the outliers, we simply add one further dimension corresponding to the range
that is not covered by the channel representation. e de-correlation cannot be solved easily in
this case, but we can assume that this additional dimension is independent of the other coeffi-
cients in the representation and therefore add the dimension directly to the de-correlated vector
in terms of
0
. e corresponding concentration parameter ˛
0
is not necessarily identical to the
other
. For notational simplicity, we however stick to the direct formulation in terms of channel
coefficient c
n
; n D 1; : : : ; N for the subsequent sections.
6.2 COMPARING CHANNEL REPRESENTATIONS
In many applications, the estimated distribution of measurements is only an intermediate step
in some processing chain. In parameter regression problems, the subsequent step is to extract
the modes of one distribution (see Chapter 5). In matching problems, two or more distributions
need to be compared to produce matching scores.
e latter is, for instance, used in tracking as suggested by Felsberg [2013], where an
appearance model is built over time, consisting of a channel coded feature map (see Chapter 4)
that represents the empirical distribution of gray values over the tracking window. If a new frame
is processed, candidate windows in the vicinity of the predicted position are to be compared
to the existing model. e candidate window with the best score is chosen as the new object
location; see Figure 6.2.
In the predecessor to the work of Felsberg [2013], Sevilla-Lara and Learned-Miller
[2012] suggest using the L
1
-distance to compare (smoothed) histograms, which is also applied
for the channel-based tracker.
Obviously, it makes much more sense to use the L
1
-distance between two channel vectors
c and c
0
d
1
.c; c
0
/ D
N
X
nD1
jc
n
c
0
n
j (6.12)
or their Hellinger distance (jcj denotes the L
1
-norm of c)
H
1=2
.c; c
0
/ D
1
2
N
X
nD1
.
p
c
n
p
c
0
n
/
2
D
jcj C jc
0
j
2
N
X
nD1
p
c
n
c
0
n
(6.13)