Var(c)=n∑i=1c2in−1=n∑i=1(w′yi)2n−1
Linearity implies that anything without an i subscript can be taken outside the summation sign.
Var(c)=n∑i=1c2in−1=w′(n∑i=1yiy′i)wn−1
The order of scalar multiplication does not matter allowing the following
Var(c)=w′⎛⎜ ⎜ ⎜⎝n∑i=1yiy′in−1⎞⎟ ⎟ ⎟⎠w=w′Sw
Recall S is the variance covariance matrix
We want to choose w to maximise the variance while ensuring that w21+…+w2p=w′w=1. We write this as
maxww′Sws.t.w′w=1
Solving the constrained optimisation above is equivalent to solving the following unconstrained problem
maxw,λw′Sw−λ(w′w−1)
Differentiating w.r.t. w gives
∂(w′Sw−λ(w′w−1))∂w=2Sw−2λw
Differentiating w.r.t. λ gives
∂(w′Sw−λ(w′w−1))∂λ=−(w′w−1)
The key result is that for any square, symmetric matrix A it holds that
∂w′Aw∂w=2Aw This is the matrix version of the rule that the derivative of ∂aw2/∂w=2aw. From this result, the matrix result can be derived (but this is tedious).
The gradient will be zero when 2Sw−2λw=0 or simplifying when
Sw=λw
This is a very famous problem known as the eigenvalue problem. Suppose ~λ and ~w provide a solutions then
The gradient will be zero when 2Sw−2λw=0 or simplifying when
Sw=λw
This is a very famous problem known as the eigenvalue problem. Suppose ~λ and ~w provide a solutions then
We have already shown that the variance will be ~w′S~w. Since ~w is an eigenvector it must hold that
S~w=~λ~w
We have already shown that the variance will be ~w′S~w. Since ~w is an eigenvector it must hold that
S~w=~λ~w which implies
~w′S~w=~w′~λ~w=~λ~w′~w
Since S is a symmetric matrix it can decomposed as
S(p×p)=W(p×p)Λ(p×p)W(p×p)′
Since S is a symmetric matrix it can decomposed as
S(p×p)=W(p×p)Λ(p×p)W(p×p)′
Since S is a symmetric matrix it can decomposed as
S(p×p)=W(p×p)Λ(p×p)W(p×p)′
Since S is a symmetric matrix it can decomposed as
S(p×p)=W(p×p)Λ(p×p)W(p×p)′
It can be shown that an equivalent way of writing the eigenvalue decomposition is
S(p×p)=W(p×p)Λ(p×p)W(p×p)′ =p∑j=1λj(1×1)wj(p×1)w′j(1×p)
If some eigenvalues are small they can be ignored.
S=p∑j=1λjwjw′j ≈r∑j=1λjwjw′j
Only r<<p eigenvalues are used.
Y(n×p)=U(n×n)D(n×p)V′(p×p)
The matrices U and V are rotations
Y=min(n,p)∑j=1djujv′j ≈r∑j=1djujv′j
for r<<min(n,p)
biplot
Keyboard shortcuts
↑, ←, Pg Up, k | Go to previous slide |
↓, →, Pg Dn, Space, j | Go to next slide |
Home | Go to first slide |
End | Go to last slide |
Number + Return | Go to specific slide |
b / m / f | Toggle blackout / mirrored / fullscreen mode |
c | Clone slideshow |
p | Toggle presenter mode |
t | Restart the presentation timer |
?, h | Toggle this help |
Esc | Back to slideshow |