Motivating Eigenvectors
- Let's consider the two-dimensional linear transformation:
- As a review, this transformation matrix says we're moving our current basis vectors and to and
-
If we observe the span of a particular vector during the transformation (to the new vector space), then most vectors on this span will get knocked off their span after the transformation
- For example, if we observe the span of the vector , then and will span this vector before the transformation
- However, those vectors , , and (in the original vector space) will most likely not span the new vector in the transformed vector space
- However, some special vectors do remain on their span after the transformation to the new vector space
- These special vectors are called eigenvectors
- Essentially, an eigenvector of a linear transformation is a vector, where the span of that vector (in the original vector space) is the same as the span of that vector after the transformation (in the new vector space)
- Roughly, we can think of eigenvectors as uncorrelated features of our data
Example of Eigenvectors
- In this specific example, the original basis vector (i.e. i-hat) is an eigenvector
- This is because the span of i-hat is the x-axis, and from the first column of the matrix, we can see that i-hat moves over to 3 times itself, which is still on that x-axis
- More specifically, any other vector on the x-axis is also just stretched by a factor of 3 because if the way linear transformations work
- Hence, any other vector on the x-axis remains on its own span
- There are many other eigenvectors of this transformation, such as
- Although the vector is an eigenvector of the linear transformation , the vector is not always an eigenvector for every linear transformation
Properties of Eigenvalues
- An eigenvalue is the factor by which it is stretched or squished during the transformation
- Roughly, we can think of an eigenvalue as the amount of information retained by the features
- Every eigenvector has an eigenvalue associated with it
-
The following are some properties associated with eigenvalues:
- When the eigenvalue is 0, then the vector disappears because there is no information about an axis after the transformation (i.e. linear dependence)
- When the eigenvalue is 0.5, then the vector gets squished by a factor of 0.5 after the transformation
- When the eigenvalue is -0.5, then the vector gets flipped and squished by a factor of 0.5 after the transformation
-
When the eigenvalue is 1, then the vector isn't stretched or squished after the transformation, which could mean the following:
- The vector gets rotated, indicating no valid eigenvector
- The vector gets sheared, indicating the vectors on the un-sheared axis are the only possible eigenvectors
- Nothing happened, indicating the eigenvector can be any vector
- When the eigenvalue is 10, then the vector gets stretched by a factor of 10 after the transformation
- When the eigenvalue is -10, then the vector gets flipped and stretched by a factor of 10 after the transformation
Purpose of Eigenvalues and Eigenvectors
- We can understand a linear transformation by looking at the matrix
- Specifically, we can read off the columns of this matrix as the landing spots for basis vectors
- However, often a better way to understand a linear transformation is by looking at its eigenvectors and eigenvalues
- This is because it's much easier to think of a transformation in terms of some axis or rotation and an angle by which it is rotating, rather than the full 3-by-3 matrix associated with that transformation
-
In other words, we typically want to understand the following associated with a linear transformation:
- How much a vector stretched or squishes
- How a vector is rotated
- If a vector is flipped
- An eigenvalue and eigenvector can help us intuitively understand the above using a single coefficient, whereas the transformation matrix requires additional calculations to help us understand the above
Notion of Eigenvectors and Eigenvalues
- Symbolically, the following formula defines the general idea behind eigenvectors:
- Where is a transformation matrix from our original vector space to a transformed vector space
- Where is an eigenvector
- Where is a eigenvalue
- Where
Intepreting the Method of Eigenvectors and EigenValues
- The expression is saying the matrix multiplication of and gives the same result as just scaling the eigenvector by some number
- Therefore, finding the eigenvectors and their eigenvalues of a transformation matrix comes down to finding the values of and that satisfy this formula
- We can also write as the following:
- Here, is the identity matrix
- This has the same effect as the previous formula, since we're scaling the vector by the same factor
- In this matrix notation, we think of the columns of matrix as representing what happens to each basis vector (i.e. scaling each basis vector by the eigenvalue )
Finding Eigenvalues using Determinants
- Before, we rearranged our equation from to
- We can rearrange our equation one more time from to the following:
- Since can be represented as a single matrix, is a transformation matrix in itself
- Now, we're looking for a non-zero eigenvector , such that the non-zero matrix multiplied by gives us the zero vector
- Therefore, the determinant of needs to equal 0 to satisfy this equation, since and both need to be non-zero
- In other words, we're looking for an eigenvalue that satisfies the following:
- Meaning, the eigenvalue that satisfies this equation will squish space into a lower dimension, since that's what a determinant of zero implies
Example of Computing Eigenvalues
- Let's say we have a transformation matrix equal to the following:
-
First, we can determine the eigenvalues of the linear transformation by finding that satisfies
- We can reformulate into the quadratic polynomial function
- We can conclude that the only possible eigenvalues are or for this linear transformation
-
Next, we can determine the eigenvectors of the linear transformation by plugging in 2 or 3 for into and solve for in
- We can reformulate into
- We can conclude that a solution for this equation is
- Therefore, all the vectors on the diagonal line spanned by
- Keep in mind, it is possible to have just one eigenvalue, but with more than just a single line of eigenvectors
Notion of Eigenbasis
- If our basis vectors are eigenvectors (i.e. and ), then our transformation matrix is called an eigenbasis
- An eigenbasis is a diagonal matrix
- When our transformation matrix is a diagonal matrix, our transformation becomes much more intuitive and computationally efficient
- We are rarely lucky enough where our basis vectors are eigenvectors
-
However, we can always choose a different set of basis vectors that are eigenvectors and span the full space (only if our transformation has enough eigenvectors in the first place)
- We can do this using the change of basis formula
- Then, we can change our coordinate system so that these eigenvectors are our basis vectors instead
-
Therefore, if we ever wanted to perform an intensive calculation on our transformation matrix (such as computing the 100th power of the transformation matrix), then it would be much easier to change to an eigenbasis, compute the intensive calculation, then convert back to our standard system
- Specifically, we can do this using the change of basis formula