Change of basis of a linear map defined by non-square matrix
Clash Royale CLAN TAG#URR8PPP
up vote
1
down vote
favorite
The given:
Let a linear map $L : V â U$ be given in the basis $(e_1, e_2, e_3)$ of $V$ and in the basis $(f_1, f_2)$ of $U$ by $beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix$. Find the matrix of $L$ with respect to the bases
$(e_1, e_1 + e_2, e_1 + e_2 + e_3)$ and $(f_1, f_1 + f_2)$.
Now I know I am being stupid in some way, but I can't make this work. I want to say: we have new bases $(e_1,e_1+e_2,e_1+e_2+e_3)$ and $(f_1,f_1+f_2)$ which correspond to $C = beginpmatrix
1 & 1 & 1 \
0 & 1 & 1 \
0 & 0 & 1
endpmatrix$ and $D = beginpmatrix
1 & 1 \
0 & 1
endpmatrix$ respectively. And Using our change of bases formula, one version of $L_1' = C^TAC$ and the other is $L_2' = D^TAD$. But here I run into a problem of dimension, the matrix multiplication does not work. I had thought perhaps to try $L'=DAC$, since this expression has workable dimension, but that's my only reason for trying it. The basis change examples I've seen before involve expressions with one matrix and either its transpose or inverse.
Apologies again for my stupidity here, and thanks very much in advance for any assistance.
Edit: Initial problem had typo on $D$, one-zero were swapped incorrectly.
linear-algebra matrices linear-transformations change-of-basis
add a comment |Â
up vote
1
down vote
favorite
The given:
Let a linear map $L : V â U$ be given in the basis $(e_1, e_2, e_3)$ of $V$ and in the basis $(f_1, f_2)$ of $U$ by $beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix$. Find the matrix of $L$ with respect to the bases
$(e_1, e_1 + e_2, e_1 + e_2 + e_3)$ and $(f_1, f_1 + f_2)$.
Now I know I am being stupid in some way, but I can't make this work. I want to say: we have new bases $(e_1,e_1+e_2,e_1+e_2+e_3)$ and $(f_1,f_1+f_2)$ which correspond to $C = beginpmatrix
1 & 1 & 1 \
0 & 1 & 1 \
0 & 0 & 1
endpmatrix$ and $D = beginpmatrix
1 & 1 \
0 & 1
endpmatrix$ respectively. And Using our change of bases formula, one version of $L_1' = C^TAC$ and the other is $L_2' = D^TAD$. But here I run into a problem of dimension, the matrix multiplication does not work. I had thought perhaps to try $L'=DAC$, since this expression has workable dimension, but that's my only reason for trying it. The basis change examples I've seen before involve expressions with one matrix and either its transpose or inverse.
Apologies again for my stupidity here, and thanks very much in advance for any assistance.
Edit: Initial problem had typo on $D$, one-zero were swapped incorrectly.
linear-algebra matrices linear-transformations change-of-basis
add a comment |Â
up vote
1
down vote
favorite
up vote
1
down vote
favorite
The given:
Let a linear map $L : V â U$ be given in the basis $(e_1, e_2, e_3)$ of $V$ and in the basis $(f_1, f_2)$ of $U$ by $beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix$. Find the matrix of $L$ with respect to the bases
$(e_1, e_1 + e_2, e_1 + e_2 + e_3)$ and $(f_1, f_1 + f_2)$.
Now I know I am being stupid in some way, but I can't make this work. I want to say: we have new bases $(e_1,e_1+e_2,e_1+e_2+e_3)$ and $(f_1,f_1+f_2)$ which correspond to $C = beginpmatrix
1 & 1 & 1 \
0 & 1 & 1 \
0 & 0 & 1
endpmatrix$ and $D = beginpmatrix
1 & 1 \
0 & 1
endpmatrix$ respectively. And Using our change of bases formula, one version of $L_1' = C^TAC$ and the other is $L_2' = D^TAD$. But here I run into a problem of dimension, the matrix multiplication does not work. I had thought perhaps to try $L'=DAC$, since this expression has workable dimension, but that's my only reason for trying it. The basis change examples I've seen before involve expressions with one matrix and either its transpose or inverse.
Apologies again for my stupidity here, and thanks very much in advance for any assistance.
Edit: Initial problem had typo on $D$, one-zero were swapped incorrectly.
linear-algebra matrices linear-transformations change-of-basis
The given:
Let a linear map $L : V â U$ be given in the basis $(e_1, e_2, e_3)$ of $V$ and in the basis $(f_1, f_2)$ of $U$ by $beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix$. Find the matrix of $L$ with respect to the bases
$(e_1, e_1 + e_2, e_1 + e_2 + e_3)$ and $(f_1, f_1 + f_2)$.
Now I know I am being stupid in some way, but I can't make this work. I want to say: we have new bases $(e_1,e_1+e_2,e_1+e_2+e_3)$ and $(f_1,f_1+f_2)$ which correspond to $C = beginpmatrix
1 & 1 & 1 \
0 & 1 & 1 \
0 & 0 & 1
endpmatrix$ and $D = beginpmatrix
1 & 1 \
0 & 1
endpmatrix$ respectively. And Using our change of bases formula, one version of $L_1' = C^TAC$ and the other is $L_2' = D^TAD$. But here I run into a problem of dimension, the matrix multiplication does not work. I had thought perhaps to try $L'=DAC$, since this expression has workable dimension, but that's my only reason for trying it. The basis change examples I've seen before involve expressions with one matrix and either its transpose or inverse.
Apologies again for my stupidity here, and thanks very much in advance for any assistance.
Edit: Initial problem had typo on $D$, one-zero were swapped incorrectly.
linear-algebra matrices linear-transformations change-of-basis
linear-algebra matrices linear-transformations change-of-basis
edited 1 hour ago
asked 1 hour ago
Raj
826616
826616
add a comment |Â
add a comment |Â
3 Answers
3
active
oldest
votes
up vote
2
down vote
TL;DR: The correct formula is $D^-1AC$.
Change of basis is not as hard as it looks and you don't need to memorize any formulas.
First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.
Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.
So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.
On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$
Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$
The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:
$$
[I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
$$
Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.
To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.
In your example, this is precisely your matrix $C$.
To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.
But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?
I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since
$$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.
Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.
Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
â Raj
8 mins ago
@Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
â Ennar
7 mins ago
Oh wow, I didn't notice your answer. Whoops. +1
â Theo Bendit
6 mins ago
1
@Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
â Ennar
3 mins ago
add a comment |Â
up vote
2
down vote
I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.
Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
$$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.
Such a matrix has the following lovely property (and is completely defined by this property):
$$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$
This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.
This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have
$$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$
and so
$$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Note also that, if $mathrmid : V to V$ is the identity operator, then
$$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$
which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then
$$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Similarly,
$$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$
What this means is
$$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$
From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then
beginalign*
[T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
&= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
endalign*
It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.
Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
beginalign*
B_1 &= (e_1, e_2, e_3) \
B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
B_2 &= (f_1, f_2) \
B_2' &= (f_1, f_1 + f_2).
endalign*
We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute
beginalign*
[L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
&= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
endalign*
We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have
$$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$
Similarly,
$$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$
and so
$$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$
Finally, this gives us,
$$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
3 & 7 & 12 endpmatrix.$$
My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
â Raj
7 mins ago
add a comment |Â
up vote
1
down vote
You should have gotten
$$
D =
left(
beginarraycc
1 & 1 \
0 & 1
endarray
right)
$$
Then
$$
D^-1A C =
left(
beginarrayccc
-3 & -6 & -9\
3 & 7 & 12
endarray
right)
$$
add a comment |Â
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
2
down vote
TL;DR: The correct formula is $D^-1AC$.
Change of basis is not as hard as it looks and you don't need to memorize any formulas.
First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.
Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.
So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.
On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$
Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$
The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:
$$
[I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
$$
Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.
To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.
In your example, this is precisely your matrix $C$.
To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.
But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?
I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since
$$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.
Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.
Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
â Raj
8 mins ago
@Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
â Ennar
7 mins ago
Oh wow, I didn't notice your answer. Whoops. +1
â Theo Bendit
6 mins ago
1
@Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
â Ennar
3 mins ago
add a comment |Â
up vote
2
down vote
TL;DR: The correct formula is $D^-1AC$.
Change of basis is not as hard as it looks and you don't need to memorize any formulas.
First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.
Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.
So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.
On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$
Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$
The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:
$$
[I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
$$
Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.
To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.
In your example, this is precisely your matrix $C$.
To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.
But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?
I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since
$$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.
Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.
Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
â Raj
8 mins ago
@Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
â Ennar
7 mins ago
Oh wow, I didn't notice your answer. Whoops. +1
â Theo Bendit
6 mins ago
1
@Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
â Ennar
3 mins ago
add a comment |Â
up vote
2
down vote
up vote
2
down vote
TL;DR: The correct formula is $D^-1AC$.
Change of basis is not as hard as it looks and you don't need to memorize any formulas.
First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.
Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.
So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.
On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$
Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$
The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:
$$
[I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
$$
Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.
To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.
In your example, this is precisely your matrix $C$.
To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.
But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?
I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since
$$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.
Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.
TL;DR: The correct formula is $D^-1AC$.
Change of basis is not as hard as it looks and you don't need to memorize any formulas.
First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.
Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.
So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.
On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$
Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$
The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:
$$
[I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
$$
Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.
To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.
In your example, this is precisely your matrix $C$.
To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.
But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?
I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since
$$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.
Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.
answered 25 mins ago
Ennar
13.4k32343
13.4k32343
Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
â Raj
8 mins ago
@Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
â Ennar
7 mins ago
Oh wow, I didn't notice your answer. Whoops. +1
â Theo Bendit
6 mins ago
1
@Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
â Ennar
3 mins ago
add a comment |Â
Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
â Raj
8 mins ago
@Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
â Ennar
7 mins ago
Oh wow, I didn't notice your answer. Whoops. +1
â Theo Bendit
6 mins ago
1
@Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
â Ennar
3 mins ago
Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
â Raj
8 mins ago
Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
â Raj
8 mins ago
@Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
â Ennar
7 mins ago
@Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
â Ennar
7 mins ago
Oh wow, I didn't notice your answer. Whoops. +1
â Theo Bendit
6 mins ago
Oh wow, I didn't notice your answer. Whoops. +1
â Theo Bendit
6 mins ago
1
1
@Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
â Ennar
3 mins ago
@Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
â Ennar
3 mins ago
add a comment |Â
up vote
2
down vote
I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.
Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
$$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.
Such a matrix has the following lovely property (and is completely defined by this property):
$$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$
This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.
This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have
$$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$
and so
$$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Note also that, if $mathrmid : V to V$ is the identity operator, then
$$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$
which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then
$$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Similarly,
$$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$
What this means is
$$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$
From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then
beginalign*
[T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
&= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
endalign*
It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.
Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
beginalign*
B_1 &= (e_1, e_2, e_3) \
B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
B_2 &= (f_1, f_2) \
B_2' &= (f_1, f_1 + f_2).
endalign*
We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute
beginalign*
[L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
&= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
endalign*
We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have
$$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$
Similarly,
$$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$
and so
$$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$
Finally, this gives us,
$$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
3 & 7 & 12 endpmatrix.$$
My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
â Raj
7 mins ago
add a comment |Â
up vote
2
down vote
I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.
Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
$$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.
Such a matrix has the following lovely property (and is completely defined by this property):
$$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$
This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.
This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have
$$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$
and so
$$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Note also that, if $mathrmid : V to V$ is the identity operator, then
$$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$
which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then
$$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Similarly,
$$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$
What this means is
$$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$
From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then
beginalign*
[T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
&= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
endalign*
It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.
Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
beginalign*
B_1 &= (e_1, e_2, e_3) \
B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
B_2 &= (f_1, f_2) \
B_2' &= (f_1, f_1 + f_2).
endalign*
We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute
beginalign*
[L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
&= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
endalign*
We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have
$$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$
Similarly,
$$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$
and so
$$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$
Finally, this gives us,
$$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
3 & 7 & 12 endpmatrix.$$
My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
â Raj
7 mins ago
add a comment |Â
up vote
2
down vote
up vote
2
down vote
I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.
Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
$$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.
Such a matrix has the following lovely property (and is completely defined by this property):
$$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$
This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.
This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have
$$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$
and so
$$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Note also that, if $mathrmid : V to V$ is the identity operator, then
$$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$
which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then
$$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Similarly,
$$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$
What this means is
$$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$
From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then
beginalign*
[T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
&= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
endalign*
It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.
Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
beginalign*
B_1 &= (e_1, e_2, e_3) \
B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
B_2 &= (f_1, f_2) \
B_2' &= (f_1, f_1 + f_2).
endalign*
We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute
beginalign*
[L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
&= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
endalign*
We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have
$$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$
Similarly,
$$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$
and so
$$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$
Finally, this gives us,
$$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
3 & 7 & 12 endpmatrix.$$
I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.
Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
$$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.
Such a matrix has the following lovely property (and is completely defined by this property):
$$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$
This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.
This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have
$$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$
and so
$$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Note also that, if $mathrmid : V to V$ is the identity operator, then
$$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$
which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then
$$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$
Similarly,
$$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$
What this means is
$$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$
From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then
beginalign*
[T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
&= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
&= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
endalign*
It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.
Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
beginalign*
B_1 &= (e_1, e_2, e_3) \
B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
B_2 &= (f_1, f_2) \
B_2' &= (f_1, f_1 + f_2).
endalign*
We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute
beginalign*
[L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
&= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
endalign*
We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have
$$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$
Similarly,
$$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$
and so
$$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$
Finally, this gives us,
$$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
3 & 7 & 12 endpmatrix.$$
answered 15 mins ago
Theo Bendit
14k12045
14k12045
My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
â Raj
7 mins ago
add a comment |Â
My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
â Raj
7 mins ago
My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
â Raj
7 mins ago
My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
â Raj
7 mins ago
add a comment |Â
up vote
1
down vote
You should have gotten
$$
D =
left(
beginarraycc
1 & 1 \
0 & 1
endarray
right)
$$
Then
$$
D^-1A C =
left(
beginarrayccc
-3 & -6 & -9\
3 & 7 & 12
endarray
right)
$$
add a comment |Â
up vote
1
down vote
You should have gotten
$$
D =
left(
beginarraycc
1 & 1 \
0 & 1
endarray
right)
$$
Then
$$
D^-1A C =
left(
beginarrayccc
-3 & -6 & -9\
3 & 7 & 12
endarray
right)
$$
add a comment |Â
up vote
1
down vote
up vote
1
down vote
You should have gotten
$$
D =
left(
beginarraycc
1 & 1 \
0 & 1
endarray
right)
$$
Then
$$
D^-1A C =
left(
beginarrayccc
-3 & -6 & -9\
3 & 7 & 12
endarray
right)
$$
You should have gotten
$$
D =
left(
beginarraycc
1 & 1 \
0 & 1
endarray
right)
$$
Then
$$
D^-1A C =
left(
beginarrayccc
-3 & -6 & -9\
3 & 7 & 12
endarray
right)
$$
answered 1 hour ago
Will Jagy
98.4k595196
98.4k595196
add a comment |Â
add a comment |Â
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2928327%2fchange-of-basis-of-a-linear-map-defined-by-non-square-matrix%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password