Hovedinnhold

## Eigen-everything

Gjeldende klokkeslett:0:00Total varighet:7:43

# Introduction to eigenvalues and eigenvectors

## Videotranskripsjon

For any transformation that maps
from Rn to Rn, we've done it implicitly, but it's been
interesting for us to find the vectors that essentially just
get scaled up by the transformations. So the vectors that have the
form-- the transformation of my vector is just equal
to some scaled-up version of a vector. And if this doesn't look
familiar, I can jog your memory a little bit. When we were looking for
basis vectors for the transformation--
let me draw it. This was from R2 to R2. So let me draw R2 right here. And let's say I had the
vector v1 was equal to the vector 1, 2. And we had the lines spanned
by that vector. We did this problem several
videos ago. And I had the transformation
that flipped across this line. So if we call that line l, T was
the transformation from R2 to R2 that flipped vectors
across this line. So it flipped vectors
across l. So if you remember that
transformation, if I had some random vector that looked like
that, let's say that's x, that's vector x, then the
transformation of x looks something like this. It's just flipped across
that line. That was the transformation
of x. And if you remember that video,
we were looking for a change of basis that would allow
us to at least figure out the matrix for the
transformation, at least in an alternate basis. And then we could figure
out the matrix for the transformation in the
standard basis. And the basis we picked were
basis vectors that didn't get changed much by the
transformation, or ones that only got scaled by the
transformation. For example, when I took the
transformation of v1, it just equaled v1. Or we could say that the
transformation of v1 just equaled 1 times v1. So if you just follow this
little format that I set up here, lambda, in this
case, would be 1. And of course, the vector
in this case is v1. The transformation just
scaled up v1 by 1. In that same problem, we had
the other vector that we also looked at. It was the vector minus-- let's
say it's the vector v2, which is-- let's say
it's 2, minus 1. And then if you take the
transformation of it, since it was orthogonal to the
line, it just got flipped over like that. And that was a pretty
interesting vector force as well, because the transformation
of v2 in this situation is equal to what? Just minus v2. It's equal to minus v2. Or you could say that the
transformation of v2 is equal to minus 1 times v2. And these were interesting
vectors for us because when we defined a new basis with these
guys as the basis vector, it was very easy to figure out
our transformation matrix. And actually, that basis was
very easy to compute with. And we'll explore that a little
bit more in the future. But hopefully you realize that
these are interesting vectors. There was also the cases where
we had the planes spanned by some vectors. And then we had another vector
that was popping out of the plane like that. And we were transforming things
by taking the mirror image across this and we're
like, well in that transformation, these red
vectors don't change at all and this guy gets
flipped over. So maybe those would make
for good bases. Or those would make for
good basis vectors. And they did. So in general, we're always
interested with the vectors that just get scaled up
by a transformation. It's not going to be
all vectors, right? This vector that I drew here,
this vector x, it doesn't just get scaled up, it actually gets
changed, this direction gets changed. The vectors that get scaled up
might switch direct-- might go from this direction to that
direction, or maybe they go from that. Maybe that's x and then the
transformation of x might be a scaled up version of x. Maybe it's that. The actual, I guess, line that
they span will not change. And so that's what we're going
to concern ourselves with. These have a special name. And they have a special name and
I want to make this very clear because they're useful. It's not just some mathematical
game we're playing, although sometimes
we do fall into that trap. But they're actually useful. They're useful for defining
bases because in those bases it's easier to find
transformation matrices. They're more natural coordinate
systems. And oftentimes, the transformation
matrices in those bases are easier to compute with. And so these have
special names. Any vector that satisfies this
right here is called an eigenvector for the
transformation T. And the lambda, the multiple
that it becomes-- this is the eigenvalue associated with
that eigenvector. So in the example I just gave
where the transformation is flipping around this line,
v1, the vector 1, 2 is an eigenvector of our
transformation. So 1, 2 is an eigenvector. And it's corresponding
eigenvalue is 1. This guy is also an
eigenvector-- the vector 2, minus 1. He's also an eigenvector. A very fancy word, but all it
means is a vector that's just scaled up by a transformation. It doesn't get changed in any
more meaningful way than just the scaling factor. And it's corresponding
eigenvalue is minus 1. If this transformation--
I don't know what its transformation matrix is. I forgot what it was. We actually figured it
out a while ago. If this transformation matrix
can be represented as a matrix vector product-- and it should
be; it's a linear transformation-- then any
v that satisfies the transformation of-- I'll say
transformation of v is equal to lambda v, which also would
be-- you know, the transformation of [? v ?] would just be A times v. These are also called
eigenvectors of A, because A is just really the matrix
representation of the transformation. So in this case, this would be
an eigenvector of A, and this would be the eigenvalue
associated with the eigenvector. So if you give me a matrix that
represents some linear transformation. You can also figure
these things out. Now the next video we're
actually going to figure out a way to figure these
things out. But what I want you to
appreciate in this video is that it's easy to say,
oh, the vectors that don't get changed much. But I want you to understand
what that means. It literally just gets scaled up
or maybe they get reversed. Their direction or the
lines they span fundamentally don't change. And the reason why they're
interesting for us is, well, one of the reasons why they're
interesting for us is that they make for interesting basis
vectors-- basis vectors whose transformation matrices
are maybe computationally more simpler, or ones that make for
better coordinate systems.