------------------------ PCA, SVD and MDS ------------------------ SVD notation and example ------------------------------ .. math:: {\bf X}_{n\times p}= U_{n\times p}S_{p\times p}V_{p\times p}', V'V=I_p, U'U=I_p. {\bf S}_{p\times p}= \left(\begin{array}{lccr} s_1 &0& \cdots& 0\\ 0 & s_2 & \cdots& 0\\ 0 & \cdots & s_{k-1}& 0\\ 0 & 0 & \cdots& s_{k}\\ \end{array} \right) where r is the rank of X. Generating a rank three matrix ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Now we have generated an ``approximately rank three`` matrix. .. rcode:: :silent: :force: rm(X) .. rcode:: :force: load("/Users/susan/sphinx/data/SVD3ex.save") svd(X3)$d res3=eigen(t(X3)%*%X3) svd(X3)$d^2 #Comparing the ouput from a PCA and an SVD? require(ade4) X3.pca=dudi.pca(X3,scannf=F,nf=4) X3.pca$eig X3.pca$co So, the output from the SVD, Eigendecomposition and PCA are not the same? Why Not? ~~~~~~~~~~~~~~ Well, for PCA the default is for the matrix to be centered by columns first, so if we don't do that, we should get the same as above, let's try it. .. rcode:: load("/Users/susan/sphinx/data/SVD3ex.save") require(ade4) X3.pca=dudi.pca(X3,scannf=F,nf=4,center=F) X3.pca$co res3$vectors[,1] Not right yet, so not only does the dudi.pca center the data but by default it also rescales it so that all the variables (columns) have variance 1, so let's try not rescaling or centering. .. rcode:: load("/Users/susan/sphinx/data/SVD3ex.save") require(ade4) X3.pca=dudi.pca(X3,scannf=F,nf=4,center=F,scale=F) X3.pca$co[,1] ##still not right, but we can remark a funny thing, ##the sums of squares of this vector are equal to: sum(X3.pca$co[,1]^2) ####we compare this to X3.pca$eig and see that by coincidence the first eigenvalue matches the norm of the component in the `co` (columns) part of `X3.pca`. In fact if we look at the normalized component which is in ``X3.pca$c1``, we have an exact match: .. rcode:: load("/Users/susan/sphinx/data/SVD3ex.save") require(ade4) X3.pca=dudi.pca(X3,scannf=F,nf=4,center=F,scale=F) X3.pca$c1 res3$vectors[,1] Now we can make the eigendecomposition match in the other direction: Start with X3 then center and scale it and check that its svd and eigenvectors correspond to the simple PCA: .. rcode:: load("/Users/susan/sphinx/data/SVD3ex.save") require(ade4) X3sc=scale(X3) apply(X3sc,2,mean) apply(X3sc,2,sd) eigen(t(X3sc)%*%X3sc) res3c=dudi.pca(X3sc,nf=6,scannf=F) res3c$c1 res3c$eig ##What's the relationship between the eigenvalues, ###there must be one (the evectors are the same) eigen(t(X3sc)%*%X3sc)$values[1]/res3c$eig[1] Summary of things we noticed ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ #. Usually PCA does an eigendecomposition of the crossproduct of the standardized, recentered matrix divided by (n-1), we call this matrix the correlation matrix. #. The square of the singular values are the eigenvalues. #. The principal components have their norm equal to the eigenvalue. #. Each eigenvalue actually gives the variance explained by that component (ie. the first eigenvalue gives the variance explained by the first component) .. rflush:: PCA and MDS(multidimensional scaling) -------------------------------------------------------- We want to compute a matrix of distances between rows of X: .. rcode:: load("/Users/susan/sphinx/data/SVD3ex.save") X3c=sweep(X3,2,apply(X3,2,mean)) B=crossprod(t(X3c)) DX2=as.matrix(dist(X3c))^2 DX2c=sweep(DX2,1,apply(DX2,1,mean)) cDX2c=sweep(DX2c,2,apply(DX2c,2,mean)) apply(cDX2c,1,mean) apply(cDX2c,2,mean) B[1:4,1:4] cDX2c[1:4,1:4] resd=eigen(-cDX2c/2) resd$values[1:4] resd$vectors[,1] ####Now we compare with the output of what is known as classical MDS (multidimensional scaling). res.mds=cmdscale(dist(X3),k=4,eig=TRUE) res.mds$points[,1] res.mds$eig[1:4] require(ade4) res.pca=dudi.pca(X3,nf=4,scannf=F,scale=F) res.pca$li[,1] ######We can also compare the eigenvalues res.mds$eig[1:3]/res.pca$eig[1:3] .. rflush:: Examples of Using MDS just on distances ----------------------------------------------- .. rcode:: ###Eckman's confusions of colors data eck=read.table("/Users/susan/sphinx/data/eckman.txt",header=TRUE) eck[1:4,1:4] require(ade4) queck=quasieuclid(as.dist(1-eck)) eck.pco=dudi.pco(queck,scannf=F,nf=2) .. rplot:: :force: eck=read.table("/Users/susan/sphinx/data/eckman.txt",header=TRUE) require(ade4) queck=quasieuclid(as.dist(1-eck)) eck.pco=dudi.pco(queck,scannf=F,nf=2) biplot(eck.pco,posi="bottomright") What is this shape we see? ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Typical ouput from PCoA and PCA type methods include horseshoes and arches if there is an underlying gradient. For a talk about this in the case of election data see `slides horsehoe, Holmes `_ and the `paper `_. .. rflush:: Summary of facts about PCA and MDS ----------------------------------------- #. If we have the orginal n by p table of data (X), then a standard PCA and a classical MDS will give the same components. This is why MDS is also called PCoA or Principal Coordinate Analysis. #. The eigenvalues are different - changed by a coefficient. #. MDS can be used on data for which we do not know coordinates, only the relative distances. References -------------- Usually it takes a whole course to go through all the multivariate methods, `here `_ is such a course. For a quick refresher on eigenvectors, there is a nice informative video about eigenvalues and eigenvectors at `Khan academy EV `_.