Singular Value Decomposition (SVD) from scratch with linear operators

Singular Value Decomposition (SVD) is definitely one of the most important tools in Linear Algebra – and it can be tricky to understand. In this log, I will be deriving SVD from scratch with linear operators and reviewing its computation and applications.

Prerequisites

Key concepts and tools:

Spectral theorem, Positive Semi-definite Operator, Self-adjoint, Normal & Isometry

If you are not familiar with these concepts, the book Linear Algebra Done Right is a great reference.

Derivation

Computation

Applications

Key points: Projection, Least Squares, Image Compression & PCA

I’d like to elaborate a bit more on the image compression section here, which gives some important insights on matrix multiplication. Recall when you were learning linear algebra for the first time – matrix multiplication is just “row-column” inner product and “filling in the blanks”. However, SVD gives us a different way to think about it – taking the outer product instead of the inner product and factoring the matrix into the weighted sum of rank-1 approximations. That’s exactly how to compress the image – truncate at the top-k singular values!

You will probably see more of these kind of decompositions if you go on to learn courses such as graph-matrix analysis. Believe me, this can be very intriguing!


评论

发表回复

您的邮箱地址不会被公开。 必填项已用 * 标注