Database Reference
In-Depth Information
Table 8.1 Example of a
session matrix of a web shop
Session 1 Session 2 Session 3 Session 4
Product 1 0
1
10
5
Product 2 1
5
1
1
Product 3 0
5
0
1
statistical models, however, entails some major computational impediments,
including intractable integrals and non-convex optimization, making a realtime
implementation very difficult. The alternative approach consists in treating all
variables as observed by assigning the value 0 to the unknown ratings. Although
this may appear somewhat helter-skelter at the first glance, it may be rationalized
by considering not visiting a product as a transaction corresponding to the lowest
possible rating. Assuming, furthermore, that many of the zero entries are due to
noise rather than intrinsic, we may put the approach on a sound footing. We will
return to this discussion in Sect. 8.5 .
Now we consider a matrix of rewards A
n p n s with n s being the current
number of sessions and n p the number of different products over all sessions
observed so far. Neither the order of sessions nor the order of products within the
sessions is taken into account.
∈ R
Example 8.1 As an example, we consider a web shop with 3 products and 4
sessions, i.e., n p ¼ 3 and n s ¼ 4. The session values are displayed in Table 8.1 .
In terms of the reward assignment described above, this means, e.g., for session
2, product 1 has been clicked, whereas products 2 and 3 have moreover been added
to the basket. In session 3, product 1 has been purchased, product 2 has been
clicked, and product 3 has been skipped.
Mathematically, the matrix factorization problems arising in CF are of the form
rn s fA
ð
;
XY
Þ:
ð 8
:
1 Þ
min
,
n p r
X
C 1 R
Y
C 2 R
The rank r is usually chosen to be considerably smaller than n p . The function
f is referred to as the cost function of the factorization and, more often than not,
is chosen to be a metric. It stipulates a notion of quality of a factorization.
The sets C 1 , C 2 determine the parameter space. In terms of our signal processing
metaphor, the factor X characterizes the source, which is restricted to be a
low-dimensional subspace, and the columns Y are the intrinsic low-dimensional
parameter vectors determining the signals given by the corresponding columns
of A .
To put it even simpler, we approximate the matrix A by the product of two
smaller matrices X and Y . The cost function stipulates a notion of “closeness,” i.e.,
distance, of two matrices. Since the rank r is typically much smaller than n p and n s ,
the representation in terms of X and Y is much more compact than an explicit
representation of the entries of A .
Search WWH ::




Custom Search