Eigen-Things

Motivating Eigenvectors

  • Let's consider the two-dimensional linear transformation:
A=[3102]A = \begin{bmatrix} 3 & 1 \cr 0 & 2 \end{bmatrix}
  • As a review, this transformation matrix says we're moving our current basis vectors [1,0][1, 0] and [0,1][0, 1] to [3,0][3, 0] and [1,2][1, 2]
  • If we observe the span of a particular vector during the transformation (to the new vector space), then most vectors on this span will get knocked off their span after the transformation

    • For example, if we observe the span of the vector [1,1][1, 1], then [2,2][2, 2] and [100,100][100, 100] will span this vector before the transformation
    • However, those vectors [1,1][1, 1], [2,2][2, 2], and [100,100][100, 100] (in the original vector space) will most likely not span the new vector [1,1][1, 1] in the transformed vector space
  • However, some special vectors do remain on their span after the transformation to the new vector space
  • These special vectors are called eigenvectors
  • Essentially, an eigenvector of a linear transformation is a vector, where the span of that vector (in the original vector space) is the same as the span of that vector after the transformation (in the new vector space)
  • Roughly, we can think of eigenvectors as uncorrelated features of our data

Example of Eigenvectors

  • In this specific example, the original basis vector (i.e. i-hat) is an eigenvector
  • This is because the span of i-hat is the x-axis, and from the first column of the matrix, we can see that i-hat moves over to 3 times itself, which is still on that x-axis
  • More specifically, any other vector on the x-axis is also just stretched by a factor of 3 because if the way linear transformations work
  • Hence, any other vector on the x-axis remains on its own span
  • There are many other eigenvectors of this transformation, such as [1,1][-1, 1]
  • Although the vector [1,0][1, 0] is an eigenvector of the linear transformation [[3,0],[1,2]][[3, 0], [1, 2]], the vector [1,0][1, 0] is not always an eigenvector for every linear transformation

Properties of Eigenvalues

  • An eigenvalue is the factor by which it is stretched or squished during the transformation
  • Roughly, we can think of an eigenvalue as the amount of information retained by the features
  • Every eigenvector has an eigenvalue associated with it
  • The following are some properties associated with eigenvalues:

    • When the eigenvalue is 0, then the vector disappears because there is no information about an axis after the transformation (i.e. linear dependence)
    • When the eigenvalue is 0.5, then the vector gets squished by a factor of 0.5 after the transformation
    • When the eigenvalue is -0.5, then the vector gets flipped and squished by a factor of 0.5 after the transformation
    • When the eigenvalue is 1, then the vector isn't stretched or squished after the transformation, which could mean the following:

      • The vector gets rotated, indicating no valid eigenvector
      • The vector gets sheared, indicating the vectors on the un-sheared axis are the only possible eigenvectors
      • Nothing happened, indicating the eigenvector can be any vector
    • When the eigenvalue is 10, then the vector gets stretched by a factor of 10 after the transformation
    • When the eigenvalue is -10, then the vector gets flipped and stretched by a factor of 10 after the transformation

Purpose of Eigenvalues and Eigenvectors

  • We can understand a linear transformation by looking at the matrix
  • Specifically, we can read off the columns of this matrix as the landing spots for basis vectors
  • However, often a better way to understand a linear transformation is by looking at its eigenvectors and eigenvalues
  • This is because it's much easier to think of a transformation in terms of some axis or rotation and an angle by which it is rotating, rather than the full 3-by-3 matrix associated with that transformation
  • In other words, we typically want to understand the following associated with a linear transformation:

    • How much a vector stretched or squishes
    • How a vector is rotated
    • If a vector is flipped
  • An eigenvalue and eigenvector can help us intuitively understand the above using a single coefficient, whereas the transformation matrix requires additional calculations to help us understand the above

Notion of Eigenvectors and Eigenvalues

  • Symbolically, the following formula defines the general idea behind eigenvectors:
Ax=λxAx = \lambda x
  • Where AA is a transformation matrix from our original vector space to a transformed vector space
  • Where xx is an eigenvector
  • Where λ\lambda is a eigenvalue
  • Where λx=b\lambda x = b

Intepreting the Method of Eigenvectors and EigenValues

  • The expression Ax=λxAx = \lambda x is saying the matrix multiplication of AA and xx gives the same result as just scaling the eigenvector xx by some number λ\lambda
  • Therefore, finding the eigenvectors and their eigenvalues of a transformation matrix AA comes down to finding the values of xx and λ\lambda that satisfy this formula
  • We can also write Ax=λxAx = \lambda x as the following:
Ax=(λI)xAx = (\lambda I)x
  • Here, II is the identity matrix
  • This has the same effect as the previous formula, since we're scaling the vector by the same factor
  • In this matrix notation, we think of the columns of matrix λI\lambda I as representing what happens to each basis vector (i.e. scaling each basis vector by the eigenvalue λ\lambda)

Finding Eigenvalues using Determinants

  • Before, we rearranged our equation from Ax=λxAx = \lambda x to Ax=(λI)xAx = (\lambda I)x
  • We can rearrange our equation one more time from Ax=(λI)xAx = (\lambda I)x to the following:
(AλI)x=0(A-\lambda I)x = 0
  • Since (AλI)(A-\lambda I) can be represented as a single matrix, (AλI)(A-\lambda I) is a transformation matrix in itself
  • Now, we're looking for a non-zero eigenvector xx, such that the non-zero matrix (AλI)(A-\lambda I) multiplied by xx gives us the zero vector
  • Therefore, the determinant of (AλI)(A-\lambda I) needs to equal 0 to satisfy this equation, since (AλI)(A-\lambda I) and xx both need to be non-zero
  • In other words, we're looking for an eigenvalue that satisfies the following:
det(AλI)=0det(A-\lambda I) = 0
  • Meaning, the eigenvalue that satisfies this equation will squish space into a lower dimension, since that's what a determinant of zero implies

Example of Computing Eigenvalues

  • Let's say we have a transformation matrix AA equal to the following:
A=[3102]A = \begin{bmatrix} 3 & 1 \cr 0 & 2 \end{bmatrix}
  • First, we can determine the eigenvalues λ\lambda of the linear transformation by finding λ\lambda that satisfies det(AλI)=0det(A-\lambda I) = 0

    • We can reformulate det(AλI)det(A-\lambda I) into the quadratic polynomial function (3λ)(2λ)=0(3-\lambda)(2-\lambda)=0
    • We can conclude that the only possible eigenvalues are λ=3\lambda=3 or λ=2\lambda=2 for this linear transformation
  • Next, we can determine the eigenvectors of the linear transformation by plugging in 2 or 3 for λ\lambda into (AλI)(A-\lambda I) and solve for xx in (AλI)x=0(A-\lambda I)x = 0

    • We can reformulate (AλI)(A-\lambda I) into x+y=0x+y=0
    • We can conclude that a solution for this equation is [1,1][-1, 1]
    • Therefore, all the vectors on the diagonal line spanned by [1,1][-1, 1]
  • Keep in mind, it is possible to have just one eigenvalue, but with more than just a single line of eigenvectors

Notion of Eigenbasis

  • If our basis vectors are eigenvectors (i.e. [1,0][-1, 0] and [2,0][2, 0]), then our transformation matrix is called an eigenbasis
  • An eigenbasis is a diagonal matrix
  • When our transformation matrix is a diagonal matrix, our transformation becomes much more intuitive and computationally efficient
  • We are rarely lucky enough where our basis vectors are eigenvectors
  • However, we can always choose a different set of basis vectors that are eigenvectors and span the full space (only if our transformation has enough eigenvectors in the first place)

    • We can do this using the change of basis formula Ax=bAx = b
  • Then, we can change our coordinate system so that these eigenvectors are our basis vectors instead
  • Therefore, if we ever wanted to perform an intensive calculation on our transformation matrix (such as computing the 100th power of the transformation matrix), then it would be much easier to change to an eigenbasis, compute the intensive calculation, then convert back to our standard system

    • Specifically, we can do this using the change of basis formula A1MAA^{-1}MA

References

Previous
Next

Change of Basis

Vectors as Functions