Here is a much better explanation of PCA: <a href="https://stats.stackexchange.com/questions/2691/making-sense-of-principal-component-analysis-eigenvectors-eigenvalues" rel="nofollow">https://stats.stackexchange.com/questions/2691/making-sense-...</a><p>The key insight that many are missing is that PCA solves a series of optimization problems, namely that reconstructing the data from the first k PCs gives the best k-dimensional approximation in terms of the squared error. Even more, this is equivalent to assuming that the data lives in a k-dimensional subspace and becomes truly high-dimensional because of normally distributed noise that spills into every direction (dimension).
Best thing I’ve ever read on PCA is Madeleine Udell’s PhD-thesis [1]. It extends PCA in many directions and shows that well-known techniques fit into the developed framework. (Was also impressed with a 138 page thesis in math that is readable as well. Quite the achievement.)<p>[1] <a href="https://people.orie.cornell.edu/mru8/doc/udell15_thesis.pdf" rel="nofollow">https://people.orie.cornell.edu/mru8/doc/udell15_thesis.pdf</a>
In the UK eating example, it would be better to examine the feature-space singular vector associated with the first singular value instead of instructing the reader to "go back and look at the data in the table". PCA has already done that work, no additional (error-prone, subjective) interpretation needed.