Biology Reference
In-Depth Information
Assumption
2
is also satisfied by construction provided the error covariance ma-
trix
is diagonal (see
Lebre 2009
). Assuming uncorrelated errors between different
variables may not necessarily hold in real-world scenarios. Nevertheless, it is not
unreasonable. Assumption
3
is difficult to verify, but it is not too restrictive if the
variables included in the data set are distinct. Then from Theorem
3.1
,aVAR(1)
process whose error covariance matrix
Σ
is diagonal can be represented by a dy-
namic Bayesian network whose arcs are identified by the nonzero elements of
A
.
For an illustration, any VAR(1) process with diagonal
Σ
where matrix
A
has the
following form (where the elements
a
ij
refer to nonzero coefficients),
Σ
⎛
⎞
a
11
a
12
0
a
21
⎝
⎠
,
A
=
00
(3.14)
0
a
32
0
can be represented by the dynamic network in Fig.
3.2
c. For instance, the non-zero
coefficient
a
12
implies the arc from
X
2
to
X
1
in Fig.
3.2
a.
3.3 Dynamic Bayesian Network Learning Algorithms
Several approaches have been covered in Chap.
2
for static Bayesian networks.
Learning a dynamic Bayesian network defining a VAR model from the given data
is a very different process and amounts to identifying the nonzero coefficients of
the auto-regressive matrix
A
. Under the homogeneity assumption (Assumption
4
in
Sect.
3.2.1
), repeated time measurements can be used to perform linear regression.
Let
k
be the number of variables under study. Then each variable
X
i
,
i
=
,...,
1
k
in
a VAR(1) process satisfies
k
j
=
1
a
ij
X
j
(
t
−
1
)+
b
i
+
ε
i
(
t
)
X
i
(
t
)=
where
ε
i
(
t
)
∼
N
(
0
,
σ
i
(
t
))
.
(3.15)
However, the classic ordinary least square estimates of the regression coefficients
a
ij
and
b
i
can be computed only when
n
k
, thus ensuring that the sample covariance
matrix has full rank. For real-world data, regularized estimators are required in most
cases.
3.3.1 Least Absolute Shrinkage and Selection Operator
The
Least Absolute Shrinkage and Selection Operator
or LASSO (
Tibshirani 1996
)
is a standard procedure, first applied to network inference by
Meinshausen and
Buhlman
(
2006
). This constrained estimation procedure tends to produce some
coefficients that are exactly zero by applying an
L
1
norm penalty to their sum.
Variable
selection
is
then
straightforward:
only
nonzero
coefficients
define
Search WWH ::
Custom Search