Database Reference
In-Depth Information
(
,
)
(
)
Regardless of the approach to estimating
P
c
I
and
P
I
, there are two problems
(
|
)
with above estimation of
P
. First, the background information
I
may include
different numbers of indexes, which requires separate estimations of the model for
different sizes of
I
. Second, when collecting training data, we cannot guarantee
enough or even available samples for a certain configuration of
c
and
I
, where the
configuration refers to a particular instantiation of the number of random variables
of
c
c
I
∪
I
and their values.
To deal with the estimation of the context model efficiently, the
P
(
|
)
is
approximated using a distribution of a set of binary random variables estimated
based on the maximum entropy (ME) principle. In this approach, an image is
represented using a
C
-dimensional vector of binary random variables, denoted
Y
c
I
t
, where the value of each variable
Y
c
is defined by
=(
,
,...,
)
Y
1
Y
2
Y
C
1ift e
c
-th image is relevant to a query
0
Y
c
=
(2.86)
otherwise
1
C
's, the data utilized by the
context modeling procedure belong to the set of vertices of a
C
-dimensional hyper-
cube. Given a set of
N
training samples, denoted
Y
1
,
Instead of being from the Cartesian product of
|
I
|
+
Y
2
,...,
Y
N
, we can estimate the
and then calculate the conditional probability
P
Y
c
|
, which is
P
(
Y
)
Y
I
1
,
Y
I
2
,...,
Y
I
|
I
|
represented as
P
(
Y
c
|
Y
I
)
in what follows. To approximate the
P
(
c
|
I
)
in Eq. (
2.82
),
the following formula is utilized
P
(
Y
c
|
Y
I
)
P
(
c
|
I
)=
(2.87)
C
v
1
P
(
Y
v
|
Y
I
)
∑
=
As the size of the concept ensemble, i.e.
C
, grows, the computational intensity
of the calculation of
P
(
|
Y
I
)
increases exponentially. Therefore, it would be more
efficient if we can directly estimate
P
Y
c
(
|
Y
I
)
based on a set of training samples.
To this end, the ME approach demonstrated in [
50
] is employed, which estimates
a conditional distribution by maximizing its Rényi entropy. Essentially, the ME
principle states that the optimal model should only respect a certain set of statistics
induced from a given training set and otherwise be as uniform as possible. The
ME approach searches for the conditional distribution
P
Y
c
(
|
Y
I
)
, with the maximum
entropy, among all the distributions which are consistent with a set of statistics
extracted from the training samples. Therefore, it can be considered as constrained
optimization, which is formulated as
Y
c
P
2
]
−
y
c
,
y
I
max
(
Y
I
=
y
I
)
P
(
Y
c
=
y
c
|
Y
I
=
y
I
)
,
(2.88)
P
(
Y
c
|
Y
I
)
∈
[
0
,
1
Search WWH ::
Custom Search