Professional Documents
Culture Documents
Dr.Aruna Malapati
BITS Pilani Asst Professor
Department of CSIS
Hyderabad Campus
BITS Pilani
Hyderabad Campus
n n
VT
m A m
n
1u1v1 2u2v2
m A +
i scalar
MATRIX A IS THE SUM OF DIFFERENT MATRICES WHICH IS
REPRESENTED AS OUTER PRODUCT OF DIFFERENT VECTORS. ui vector
vi vector
BITS Pilani, Hyderabad Campus
SVD - Properties
It is always possible to decompose a real
matrix A into A = U VT , where
U, , V: unique
U, V: column orthonormal
UT U = I; VT V = I (I: identity matrix)
(Columns are orthogonal unit vectors)
: diagonal
Entries (singular values) are positive,
and sorted in decreasing order (1 2 ...
0)
BITS Pilani, Hyderabad Campus
SVD Example: Users-to-Movies
Amelie
Matrix
Alien
n
1 1 1 0 0
3 3 3 0 0
4 4 4 0 0 VT
5 5 5 0 0 = m
0 2 0 4 4
0 0 0 5 5
0 1 0 2 2 U Concepts
AKA Latent dimensions
AKA Latent factors
Amelie
Matrix
Alien
Amelie
Matrix
comedy-concept
Alien
Amelie
Matrix
SciFi-concept comedy-concept
Alien
Amelie
Matrix
SciFi-concept
Alien
Amelie
Matrix
Movie 2 rating
first right
singular vector
v1
Movie 1 rating
Movie 2 rating
first right
singular
vector
v1
Movie 1 rating
Movie 2 rating
V: movie-to-concept matrix first right
U: user-to-concept matrix singular
vector
v1
Movie 2 rating
first right
singular
variance (spread) vector
on the v1 axis
v1
Movie 2 rating
U : Gives the coordinates first right
of the points in the singular
vector
projection axis
v1
1 1 1 0 0 Movie 1 rating
Projection of users
3 3 3 0 0 on the Sci-Fi axis
4 4 4 0 0 (U ) T: 1.61 0.19 -0.01
5 5 5 0 0 5.08 0.66 -0.03
0 2 0 4 4 6.82 0.85 -0.05
0 0 0 5 5 8.43 1.04 -0.06
0 1 0 2 2 1.86 -5.60 0.84
0.86 -6.93 -0.87
0.86 -2.75 0.41
BITS Pilani, Hyderabad Campus
SVD - Interpretation #2
More details
Q: How exactly is dim. reduction done?
1 1 1 0 0 0.13 0.02
3 3 3 0 0 0.41 0.07
4 4 4 0 0 0.55 0.09 12.4 0
5 5 5 0 0 0.68 0.11 x 0 9.5 x
0 2 0 4 4 0.15 -0.59
0 0 0 5 5 0.07 -0.73
0 1 0 2 2 0.07 -0.29 0.56 0.59 0.56 0.09 0.09
0.12 -0.02 0.12 -0.69 -0.69
Frobenius norm:
A-BF = ij (Aij-Bij)2
MF = ij Mij2 is small
BITS Pilani, Hyderabad Campus
SVD Best Low Rank Approx
Sigma
A =
U
VT
B is best approximation of A
Sigma
B = U
VT
U: user-to-concept similarities
V: movie-to-concept similarities
Dimensionality reduction:
Amelie
Matrix
Alien
Alien
Casablanca
q
Serenity
Amelie
Matrix
Alien
q= 5 0 0 0 0 v2
v1
Project into concept space:
Matrix
Inner product with each
concept vector vi
Alien
Casablanca
q
Serenity
Amelie
Matrix
Alien
q= 5 0 0 0 0 v2
v1 q*v1
Project into concept space:
Matrix
Inner product with each
concept vector vi
E.g.:
Casablanca
SciFi-concept
Serenity
Amelie
Matrix
0.56 0.12
Alien
0.59 -0.02
q= 5 0 0 0 0 x 0.56 0.12 = 2.8 0.6
0.09 -0.69
0.09 -0.69
movie-to-concept
similarities (V)
SciFi-concept
Serenity
Amelie
Matrix
0.56 0.12
Alien
0.59 -0.02
q= 0 4 5 0 0 x 0.56 0.12 = 5.2 0.4
0.09 -0.69
0.09 -0.69
movie-to-concept
similarities (V)
Amelie
Matrix
SciFi-concept
Alien
d = 0 4 5 0 0 5.2 0.4
q = 5 0 0 0 0 2.8 0.6
Zero ratings in common Similarity 0
BITS Pilani, Hyderabad Campus
SVD: Drawbacks
+ Optimal low-rank approximation
in terms of Frobenius norm
- Interpretability problem:
A singular vector specifies a linear combination
of all input columns or rows
- Lack of sparsity:
Singular vectors are dense!
VT
=
U
The transpose of A is
Compute AAT
A = U VT