Biomedical Engineering Reference
In-Depth Information
is
significantly smaller than the total number of coefficients (
N
k
)), then one can use
techniques of sparse regression to find a reliable solution to (12.3).
If the model is sparse (i.e., the number of nonzero coefficients in
{
a
ij
(
t
)
}
12.2.2 Sparse Regression
Consider a multivariate linear regression model of the form
Y
=
ZW
×
×
where
Y
is a known
n
1
1 response vector,
Z
is a known
n
1
n
2
regressor matrix,
×
and
W
is the unknown model vector of size
n
2
1tobedeterminedusingthe
response
Y
and regressor
Z
. Such a problem is typically solved with techniques
such as ordinary least square regression, ridge regression, or subset selection [21].
For these techniques to work, a general requirement is that
n
1
>>
n
2
. However,
recent research shows that if
W
is sparse then it may be recovered even if
n
2
>
n
1
using the lasso regression [6, 18].
The lasso regression [18] solves the problem
2
2
||
−
||
mi
W
Y
ZW
(12.6)
||
||
≤
s.t.
W
t
(12.7)
1
2
||
||
where
2
represents the sum of squares of the coefficients of
X
(square of L2
norm of
X
)and
X
||
||
1
represents the sum of absolute values of the coefficients of
X
(its L1 norm). The parameter
t
is the regression parameter usually chosen after
cross-validation. In the extreme case, when
t
is infinitesimally small, the optimal
solution
W
corresponds to the regressor (column of
Z
) which is ''closest to''
Y
.In
the other extreme when
t
is sufficiently large, (12.7) ceases to be a constraint and
the optimal solution
W
corresponds to classical the least square solution of (12.3).
It can be verified that for any
t
,thereexistsal such that the program (12.6),
(12.7) is equivalent to the following optimization problem:
X
2
2
||
−
||
|
|
mi
W
Y
ZW
+
l
W
(12.8)
1
The programs (12.6), (12.7), and (12.8) can be solved efficiently using a tech-
nique called least angle regression [6].
12.2.3 Solving Multivariate Autoregressive Model Using Lasso
The estimation of multivariate autoregressive coefficients in (12.3) may be viewed
as a regression problem where
Y
is the response variable,
Z
is the matrix containing
the regressors, and
W
is the model to be determined. In this case, the maximum
likelihood estimate of (12.4) becomes the least square solution to the regression
problem. The lasso formulation thus becomes
Search WWH ::
Custom Search