PCA in thousands of dimensions

asked 2017-03-22 18:48:29 -0600

lovaj gravatar image

updated 2017-03-22 18:50:37 -0600

I have vectors in 4096 VLAD codes, each one of them representing an image.

I have to run PCA on them without reducing their dimension on learning datasets with less than 4096 images (e.g., the Holiday dataset with <2k images.) and obtaining the rotation matrix A.

In this paper (which also explains why I need to run this PCA without dimensionality reduction) they solved the problem with this approach:

For efficient computation of A, we compute at most the first 1,024 eigenvectors by eigendecomposition of the covariance matrix, and the remaining orthogonal complements up to D-dimensional are filled using Gram-Schmidt orthogonalization.

Now, how do I implement this using C++ libraries? I'm using OpenCV, but cv::PCA seems not to offer such a strategy. Is there any way to do this?

edit retag flag offensive close merge delete

Comments

1

answer is in your question? "we compute at most the first 1,024 eigenvectors by eigendecomposition of the covariance matrix"

use eigen lib #include < Eigen / Eigenvalues >

LBerger gravatar imageLBerger ( 2017-03-23 03:38:29 -0600 )edit

@LBerger thanks for your comment. However, I don't fully understand this. How do you obtain the covariance matrix? How do you compute the first 1024 eigenvectors? And finally, what about the D - 1024 remaining components?

lovaj gravatar imagelovaj ( 2017-03-23 04:43:39 -0600 )edit

@StevenPuttermans thanks for the reference, but how this could help me? No gram-schmidt method is cited there, neither how to compute PCA in thousands of dimensions :)

lovaj gravatar imagelovaj ( 2017-03-27 15:42:53 -0600 )edit