Biomedical Engineering Reference
In-Depth Information
K
Hx
k
,
ʛ
−
1
,
ʦ
−
1
p
(
y
)
=
N(
y
k
|
)N(
x
k
|
0
)
d
x
k
.
(B.28)
k
=
1
This
p
(
y
)
2
exp
y
k
K
1
|
ʣ
y
|
1
2
y
k
ʣ
−
1
(
)
∝
−
,
p
y
(B.29)
y
1
/
k
=
1
where
ʣ
y
is expressed as
ʣ
y
=
ʛ
−
1
ʦ
−
1
H
T
+
H
.
(B.30)
This
ʣ
y
is referred to as the model data covariance. Taking the normalization into
account, we get the marginal distribution
p
(
y
)
, such that
K
p
(
y
)
=
p
(
y
k
)
where
p
(
y
k
)
=
N(
y
k
|
0
,
ʣ
y
).
(B.31)
k
=
1
Thus,
ʣ
y
, the model data covariance, is the covariance matrix of the marginal dis-
tribution
p
(
y
k
)
.
B.5
Expectation Maximization (EM) Algorithm
B.5.1
Marginal Likelihood Maximization
The Bayesian estimate of the unknown
x
,
¯
x
k
, is computed using Eq. (
B.25
). How-
¯
ever, when computing
x
k
,
ʦ
and
ʛ
are needed. The precision matrix of the prior
ʦ
ʛ
distribution
are called the hyperparameters. Quite often, the
hyperparameters are unknown, and should also be estimated from the observation
y
.
The hyperparameters may be estimated by maximizing
p
and that of the noise
(
y
|
ʦ
,
ʛ
)
, which is the like-
lihood with respect to
is called the marginal likelihood or
the data evidence to discriminate it from the conventional likelihood
p
ʦ
and
ʛ
.This
p
(
y
|
ʦ
,
ʛ
)
(
y
|
x
)
.
using
1
The marginal likelihood can be computed with
p
(
x
|
ʦ
)
and
p
(
y
|
x
,
ʛ
)
∞
p
(
y
|
ʦ
,
ʛ
)
=
p
(
y
|
x
,
ʛ
)
p
(
x
|
ʦ
)
d
x
.
(B.32)
−∞
However, computation of
is generally quite trou-
ʦ
and
ʛ
that maximize
p
(
y
|
ʦ
,
ʛ
)
1
The notation d
x
indicates d
x
1
d
x
2