Change of basis of a linear map defined by non-square matrix

The name of the pictureThe name of the pictureThe name of the pictureClash Royale CLAN TAG#URR8PPP











up vote
1
down vote

favorite












The given:




Let a linear map $L : V → U$ be given in the basis $(e_1, e_2, e_3)$ of $V$ and in the basis $(f_1, f_2)$ of $U$ by $beginpmatrix 0 & 1 & 2 \
3 & 4 & 5 endpmatrix$
. Find the matrix of $L$ with respect to the bases
$(e_1, e_1 + e_2, e_1 + e_2 + e_3)$ and $(f_1, f_1 + f_2)$.




Now I know I am being stupid in some way, but I can't make this work. I want to say: we have new bases $(e_1,e_1+e_2,e_1+e_2+e_3)$ and $(f_1,f_1+f_2)$ which correspond to $C = beginpmatrix
1 & 1 & 1 \
0 & 1 & 1 \
0 & 0 & 1
endpmatrix$
and $D = beginpmatrix
1 & 1 \
0 & 1
endpmatrix$
respectively. And Using our change of bases formula, one version of $L_1' = C^TAC$ and the other is $L_2' = D^TAD$. But here I run into a problem of dimension, the matrix multiplication does not work. I had thought perhaps to try $L'=DAC$, since this expression has workable dimension, but that's my only reason for trying it. The basis change examples I've seen before involve expressions with one matrix and either its transpose or inverse.



Apologies again for my stupidity here, and thanks very much in advance for any assistance.



Edit: Initial problem had typo on $D$, one-zero were swapped incorrectly.










share|cite|improve this question



























    up vote
    1
    down vote

    favorite












    The given:




    Let a linear map $L : V → U$ be given in the basis $(e_1, e_2, e_3)$ of $V$ and in the basis $(f_1, f_2)$ of $U$ by $beginpmatrix 0 & 1 & 2 \
    3 & 4 & 5 endpmatrix$
    . Find the matrix of $L$ with respect to the bases
    $(e_1, e_1 + e_2, e_1 + e_2 + e_3)$ and $(f_1, f_1 + f_2)$.




    Now I know I am being stupid in some way, but I can't make this work. I want to say: we have new bases $(e_1,e_1+e_2,e_1+e_2+e_3)$ and $(f_1,f_1+f_2)$ which correspond to $C = beginpmatrix
    1 & 1 & 1 \
    0 & 1 & 1 \
    0 & 0 & 1
    endpmatrix$
    and $D = beginpmatrix
    1 & 1 \
    0 & 1
    endpmatrix$
    respectively. And Using our change of bases formula, one version of $L_1' = C^TAC$ and the other is $L_2' = D^TAD$. But here I run into a problem of dimension, the matrix multiplication does not work. I had thought perhaps to try $L'=DAC$, since this expression has workable dimension, but that's my only reason for trying it. The basis change examples I've seen before involve expressions with one matrix and either its transpose or inverse.



    Apologies again for my stupidity here, and thanks very much in advance for any assistance.



    Edit: Initial problem had typo on $D$, one-zero were swapped incorrectly.










    share|cite|improve this question

























      up vote
      1
      down vote

      favorite









      up vote
      1
      down vote

      favorite











      The given:




      Let a linear map $L : V → U$ be given in the basis $(e_1, e_2, e_3)$ of $V$ and in the basis $(f_1, f_2)$ of $U$ by $beginpmatrix 0 & 1 & 2 \
      3 & 4 & 5 endpmatrix$
      . Find the matrix of $L$ with respect to the bases
      $(e_1, e_1 + e_2, e_1 + e_2 + e_3)$ and $(f_1, f_1 + f_2)$.




      Now I know I am being stupid in some way, but I can't make this work. I want to say: we have new bases $(e_1,e_1+e_2,e_1+e_2+e_3)$ and $(f_1,f_1+f_2)$ which correspond to $C = beginpmatrix
      1 & 1 & 1 \
      0 & 1 & 1 \
      0 & 0 & 1
      endpmatrix$
      and $D = beginpmatrix
      1 & 1 \
      0 & 1
      endpmatrix$
      respectively. And Using our change of bases formula, one version of $L_1' = C^TAC$ and the other is $L_2' = D^TAD$. But here I run into a problem of dimension, the matrix multiplication does not work. I had thought perhaps to try $L'=DAC$, since this expression has workable dimension, but that's my only reason for trying it. The basis change examples I've seen before involve expressions with one matrix and either its transpose or inverse.



      Apologies again for my stupidity here, and thanks very much in advance for any assistance.



      Edit: Initial problem had typo on $D$, one-zero were swapped incorrectly.










      share|cite|improve this question















      The given:




      Let a linear map $L : V → U$ be given in the basis $(e_1, e_2, e_3)$ of $V$ and in the basis $(f_1, f_2)$ of $U$ by $beginpmatrix 0 & 1 & 2 \
      3 & 4 & 5 endpmatrix$
      . Find the matrix of $L$ with respect to the bases
      $(e_1, e_1 + e_2, e_1 + e_2 + e_3)$ and $(f_1, f_1 + f_2)$.




      Now I know I am being stupid in some way, but I can't make this work. I want to say: we have new bases $(e_1,e_1+e_2,e_1+e_2+e_3)$ and $(f_1,f_1+f_2)$ which correspond to $C = beginpmatrix
      1 & 1 & 1 \
      0 & 1 & 1 \
      0 & 0 & 1
      endpmatrix$
      and $D = beginpmatrix
      1 & 1 \
      0 & 1
      endpmatrix$
      respectively. And Using our change of bases formula, one version of $L_1' = C^TAC$ and the other is $L_2' = D^TAD$. But here I run into a problem of dimension, the matrix multiplication does not work. I had thought perhaps to try $L'=DAC$, since this expression has workable dimension, but that's my only reason for trying it. The basis change examples I've seen before involve expressions with one matrix and either its transpose or inverse.



      Apologies again for my stupidity here, and thanks very much in advance for any assistance.



      Edit: Initial problem had typo on $D$, one-zero were swapped incorrectly.







      linear-algebra matrices linear-transformations change-of-basis






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited 1 hour ago

























      asked 1 hour ago









      Raj

      826616




      826616




















          3 Answers
          3






          active

          oldest

          votes

















          up vote
          2
          down vote













          TL;DR: The correct formula is $D^-1AC$.




          Change of basis is not as hard as it looks and you don't need to memorize any formulas.



          First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.



          Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.



          So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.



          On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$



          Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$



          The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:



          $$
          [I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
          $$




          Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.



          To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.



          In your example, this is precisely your matrix $C$.



          To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.



          But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?



          I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since



          $$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.



          Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.






          share|cite|improve this answer




















          • Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
            – Raj
            8 mins ago










          • @Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
            – Ennar
            7 mins ago










          • Oh wow, I didn't notice your answer. Whoops. +1
            – Theo Bendit
            6 mins ago






          • 1




            @Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
            – Ennar
            3 mins ago


















          up vote
          2
          down vote













          I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.



          Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
          $$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
          where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.



          Such a matrix has the following lovely property (and is completely defined by this property):




          $$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$




          This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.



          This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have



          $$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$



          and so




          $$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$




          Note also that, if $mathrmid : V to V$ is the identity operator, then



          $$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$



          which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then



          $$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$



          Similarly,



          $$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$



          What this means is




          $$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$




          From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then




          beginalign*
          [T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
          &= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
          &= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
          &= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
          endalign*




          It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.



          Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
          beginalign*
          B_1 &= (e_1, e_2, e_3) \
          B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
          B_2 &= (f_1, f_2) \
          B_2' &= (f_1, f_1 + f_2).
          endalign*

          We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute



          beginalign*
          [L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
          &= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
          endalign*



          We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have



          $$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$



          Similarly,



          $$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$



          and so



          $$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$



          Finally, this gives us,



          $$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
          3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
          3 & 7 & 12 endpmatrix.$$






          share|cite|improve this answer




















          • My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
            – Raj
            7 mins ago

















          up vote
          1
          down vote













          You should have gotten
          $$
          D =
          left(
          beginarraycc
          1 & 1 \
          0 & 1
          endarray
          right)
          $$

          Then
          $$
          D^-1A C =
          left(
          beginarrayccc
          -3 & -6 & -9\
          3 & 7 & 12
          endarray
          right)
          $$






          share|cite|improve this answer




















            Your Answer




            StackExchange.ifUsing("editor", function ()
            return StackExchange.using("mathjaxEditing", function ()
            StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
            StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
            );
            );
            , "mathjax-editing");

            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "69"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            convertImagesToLinks: true,
            noModals: false,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            noCode: true, onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













             

            draft saved


            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2928327%2fchange-of-basis-of-a-linear-map-defined-by-non-square-matrix%23new-answer', 'question_page');

            );

            Post as a guest






























            3 Answers
            3






            active

            oldest

            votes








            3 Answers
            3






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes








            up vote
            2
            down vote













            TL;DR: The correct formula is $D^-1AC$.




            Change of basis is not as hard as it looks and you don't need to memorize any formulas.



            First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.



            Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.



            So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.



            On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$



            Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$



            The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:



            $$
            [I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
            $$




            Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.



            To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.



            In your example, this is precisely your matrix $C$.



            To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.



            But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?



            I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since



            $$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.



            Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.






            share|cite|improve this answer




















            • Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
              – Raj
              8 mins ago










            • @Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
              – Ennar
              7 mins ago










            • Oh wow, I didn't notice your answer. Whoops. +1
              – Theo Bendit
              6 mins ago






            • 1




              @Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
              – Ennar
              3 mins ago















            up vote
            2
            down vote













            TL;DR: The correct formula is $D^-1AC$.




            Change of basis is not as hard as it looks and you don't need to memorize any formulas.



            First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.



            Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.



            So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.



            On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$



            Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$



            The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:



            $$
            [I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
            $$




            Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.



            To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.



            In your example, this is precisely your matrix $C$.



            To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.



            But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?



            I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since



            $$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.



            Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.






            share|cite|improve this answer




















            • Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
              – Raj
              8 mins ago










            • @Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
              – Ennar
              7 mins ago










            • Oh wow, I didn't notice your answer. Whoops. +1
              – Theo Bendit
              6 mins ago






            • 1




              @Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
              – Ennar
              3 mins ago













            up vote
            2
            down vote










            up vote
            2
            down vote









            TL;DR: The correct formula is $D^-1AC$.




            Change of basis is not as hard as it looks and you don't need to memorize any formulas.



            First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.



            Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.



            So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.



            On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$



            Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$



            The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:



            $$
            [I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
            $$




            Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.



            To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.



            In your example, this is precisely your matrix $C$.



            To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.



            But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?



            I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since



            $$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.



            Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.






            share|cite|improve this answer












            TL;DR: The correct formula is $D^-1AC$.




            Change of basis is not as hard as it looks and you don't need to memorize any formulas.



            First, if $e$ is a basis for $V$ and $f$ is a basis for $U$, I will denote by $[L]^e_f$ the matrix of $Lcolon Vto U$ written in a pair of bases $(e,f)$. This works as follows: $[L]^e_f$ knows how to deal with vectors written in basis $e$ and spits out a vector written in basis $f$, i.e. $[L]^e_f[v]_e = [Lv]_f$.



            Now, we want to change bases, i.e. we want $[L]^e'_f'$. Of course, that matrix can only deal with vectors written in basis $e'$ and will spit out a vector in basis $f'$. Unfortunately, we have no idea how $[L]^e'_f'$ is supposed to look like, but fortunately, there is an easy trick which comes from a simple observation: $L = I_ULI_V,$ where $I$'s are appropriate identity operators.



            So, how does this help us? Well, we want to turn the last expression in a convenient matrix form, LHS should be $[L]^e'_f'$ and the $L$ on RHS should be $[L]^e_f$. That is, we want something like $[L]^e'_f' = [I_U]^?_? [L]^e_f [I_V]^?_?$.



            On LHS, we want to feed the matrix a vector written in basis $e'$, so $[I_V]^?_?$ should recognize it as well. Also, $[I_V]^?_?$ should give us a vector written in basis $e$, so $[L]^e_f$ could recognize it. That is, we should have $[I_V]^e'_e.$



            Since $[L]^e_f$ returns a vector written in basis $f$, $[I_U]^?_?$ should recognize it, and since $[L]^e'_f'$ should return a vector written in basis $f'$, the same applies to $[I_U]^?_?$. That is, we need $[I_U]^f_f'.$



            The whole formula now looks like $[L]^e'_f' = [I_U]^f_f' [L]^e_f [I_V]^e'_e$. We can quickly check it:



            $$
            [I_U]^f_f' [L]^e_f [I_V]^e'_e[v]_e' = [I_U]^f_f' [L]^e_f [I_Vv]_e = [I_U]^f_f' [LI_Vv]_f = [I_ULI_Vv]_f' = [Lv]_f' = [L]^e'_f'[v]_e'.
            $$




            Now, that we have a formula (that we, hopefully, now understand how it works), all it remains is to see how to get $[I_V]^e'_e$ and $[I_U]^f_f'$.



            To get $i$-th column of $[I_V]^e'_e$, we need to multiply $[I_V]^e'_e$ by column vector $[0, 0, ldots, 0, 1, 0, ldots, 0]^t$ with $1$ on the $i$-th place. But since $[I_V]^e'_e$ only accepts vectors written in basis $e'$, that column vector is $e'_i$ written in basis $e'$. All in all, the $i$-th column is $[I_V]^e'_e[e'_i]_e' = [I_Ve'_i]_e = [e'_i]_e,$ i.e., $e'_i$ written in basis $e$.



            In your example, this is precisely your matrix $C$.



            To get $[I_U]^f_f'$, you would do precisely the same, write basis vectors $f_i$ in basis $f'$. Easy, right? Ok, ok, it's not as easy since in your example we know how to write $f'_i$ in basis $f$, but how to do the opposite is not obvious.



            But there's a trick. Actually, $[I_U]^f'_f$ is quite easy, that's your matrix $D$. So, how do we get $[I_U]^f_f'$?



            I claim that $[I_U]^f_f' = ([I_U]^f'_f)^-1.$ We can see this since



            $$[I_U]^f_f'[I_U]^f'_f[u]_f' = [I_U]^f_f'[u]_f = [u]_f'$$ and thus, $[I_U]^f_f'[I_U]^f'_f$ is the identity matrix.



            Hence, in your example, $[I_U]^f_f' = D^-1$ and the correct formula is $D^-1AC$.







            share|cite|improve this answer












            share|cite|improve this answer



            share|cite|improve this answer










            answered 25 mins ago









            Ennar

            13.4k32343




            13.4k32343











            • Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
              – Raj
              8 mins ago










            • @Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
              – Ennar
              7 mins ago










            • Oh wow, I didn't notice your answer. Whoops. +1
              – Theo Bendit
              6 mins ago






            • 1




              @Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
              – Ennar
              3 mins ago

















            • Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
              – Raj
              8 mins ago










            • @Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
              – Ennar
              7 mins ago










            • Oh wow, I didn't notice your answer. Whoops. +1
              – Theo Bendit
              6 mins ago






            • 1




              @Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
              – Ennar
              3 mins ago
















            Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
            – Raj
            8 mins ago




            Thank you very much for that exceptionally detailed explanation, I'm going to digest it for a while, but I greatly appreciate the efforts.
            – Raj
            8 mins ago












            @Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
            – Ennar
            7 mins ago




            @Raj, you are welcome. Definitely also check the answer by Theo Bendit, it's very similar but I think perhaps has some details better explained.
            – Ennar
            7 mins ago












            Oh wow, I didn't notice your answer. Whoops. +1
            – Theo Bendit
            6 mins ago




            Oh wow, I didn't notice your answer. Whoops. +1
            – Theo Bendit
            6 mins ago




            1




            1




            @Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
            – Ennar
            3 mins ago





            @Theo, +1 to you as well. The time difference for such a lenghty answers is too small, so it's understandable. And I like your answer, also.
            – Ennar
            3 mins ago











            up vote
            2
            down vote













            I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.



            Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
            $$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
            where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.



            Such a matrix has the following lovely property (and is completely defined by this property):




            $$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$




            This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.



            This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have



            $$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$



            and so




            $$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$




            Note also that, if $mathrmid : V to V$ is the identity operator, then



            $$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$



            which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then



            $$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$



            Similarly,



            $$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$



            What this means is




            $$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$




            From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then




            beginalign*
            [T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
            &= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            &= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            &= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            endalign*




            It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.



            Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
            beginalign*
            B_1 &= (e_1, e_2, e_3) \
            B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
            B_2 &= (f_1, f_2) \
            B_2' &= (f_1, f_1 + f_2).
            endalign*

            We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute



            beginalign*
            [L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
            &= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
            endalign*



            We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have



            $$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$



            Similarly,



            $$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$



            and so



            $$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$



            Finally, this gives us,



            $$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
            3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
            3 & 7 & 12 endpmatrix.$$






            share|cite|improve this answer




















            • My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
              – Raj
              7 mins ago














            up vote
            2
            down vote













            I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.



            Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
            $$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
            where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.



            Such a matrix has the following lovely property (and is completely defined by this property):




            $$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$




            This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.



            This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have



            $$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$



            and so




            $$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$




            Note also that, if $mathrmid : V to V$ is the identity operator, then



            $$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$



            which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then



            $$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$



            Similarly,



            $$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$



            What this means is




            $$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$




            From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then




            beginalign*
            [T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
            &= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            &= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            &= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            endalign*




            It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.



            Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
            beginalign*
            B_1 &= (e_1, e_2, e_3) \
            B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
            B_2 &= (f_1, f_2) \
            B_2' &= (f_1, f_1 + f_2).
            endalign*

            We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute



            beginalign*
            [L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
            &= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
            endalign*



            We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have



            $$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$



            Similarly,



            $$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$



            and so



            $$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$



            Finally, this gives us,



            $$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
            3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
            3 & 7 & 12 endpmatrix.$$






            share|cite|improve this answer




















            • My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
              – Raj
              7 mins ago












            up vote
            2
            down vote










            up vote
            2
            down vote









            I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.



            Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
            $$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
            where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.



            Such a matrix has the following lovely property (and is completely defined by this property):




            $$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$




            This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.



            This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have



            $$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$



            and so




            $$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$




            Note also that, if $mathrmid : V to V$ is the identity operator, then



            $$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$



            which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then



            $$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$



            Similarly,



            $$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$



            What this means is




            $$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$




            From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then




            beginalign*
            [T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
            &= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            &= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            &= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            endalign*




            It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.



            Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
            beginalign*
            B_1 &= (e_1, e_2, e_3) \
            B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
            B_2 &= (f_1, f_2) \
            B_2' &= (f_1, f_1 + f_2).
            endalign*

            We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute



            beginalign*
            [L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
            &= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
            endalign*



            We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have



            $$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$



            Similarly,



            $$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$



            and so



            $$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$



            Finally, this gives us,



            $$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
            3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
            3 & 7 & 12 endpmatrix.$$






            share|cite|improve this answer












            I hate the change of basis formula. I think it confuses way too many people, and obscures the simple intuition going on behind the scenes.



            Recall the definition of matrices for a linear map $T : V to W$. If $B_1 = (v_1, ldots, v_m)$ is a basis for $V$ and $B_2$ is a basis for $W$ (also ordered and finite), then we define
            $$[T]_B_2 leftarrow B_1 = left([Tv_1]_B_2 mid [Tv_2]_B_2 | ldots | , [Tv_n]_B_2 right),$$
            where $[w]_B_2$ refers to the coordinate column vector of $w in W$ with respect to the basis $B_2$. Essentially, it's the matrix you get by transforming the basis $B_1$, writing the resulting vectors in terms of $B_2$, and writing the resulting coordinate vectors as columns.



            Such a matrix has the following lovely property (and is completely defined by this property):




            $$[T]_B_2 leftarrow B_1 [v]_B_1 = [Tv]_B_2.$$




            This is what makes the matrix useful. When we compute with finite-dimensional vector spaces, we tend to store vectors in terms of their coordinate vector with respect to a basis. So, this matrix allows us to directly apply $T$ to such a coordinate vector to return a coordinate vector in terms of the basis on the codomain.



            This also means that, if we also have $S : W to U$, and $U$ has a (finite, ordered) basis $B_3$, then we have



            $$[S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1[v]_B_1 = [S]_B_3 leftarrow B_2[Tv]_B_2 = [STv]_B_3,$$



            and so




            $$[ST]_B_3 leftarrow B_1 = [S]_B_3 leftarrow B_2[T]_B_2 leftarrow B_1.$$




            Note also that, if $mathrmid : V to V$ is the identity operator, then



            $$[mathrmid]_B_1 leftarrow B_1[v]_B_1 = [v]_B_1,$$



            which implies $[mathrmid]_B_1 leftarrow B_1$ is the $n times n$ identity matrix $I_n$. Moreover, if $T$ is invertible, then $operatornamedim W = n$ and then



            $$I_n = [mathrmid]_B_1 leftarrow B_1 = [T^-1T]_B_1 leftarrow B_1 = [T^-1]_B_1 leftarrow B_2[T]_B_2 leftarrow B_1.$$



            Similarly,



            $$I_n = [mathrmid]_B_2 leftarrow B_2 = [TT^-1]_B_1 leftarrow B_1 = [T]_B_2 leftarrow B_1[T^-1]_B_1 leftarrow B_2.$$



            What this means is




            $$[T]_B_2 leftarrow B_1^-1 = [T^-1]_B_1 leftarrow B_2$$




            From this, we can derive the change of basis formula. If we have a linear operator $T : V to V$ and two bases $B_1$ and $B_2$ on $V$, then




            beginalign*
            [T]_B_2 leftarrow B_2 &= [mathrmid circ T circ mathrmid]_B_2 leftarrow B_2 \
            &= [mathrmid]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            &= [mathrmid^-1]_B_2 leftarrow B_1 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            &= [mathrmid]^-1_B_1 leftarrow B_2 [T]_B_1 leftarrow B_1 [mathrmid]_B_1 leftarrow B_2 \
            endalign*




            It's easy to see that, if $B_1$ is the standard basis for $V = mathbbF^n$, then $[mathrmid]_B_1 leftarrow B_2$ is the result of putting the basis vectors in $B_2$ into columns of a matrix, and this particular case is the change of basis formula.



            Now, this works for an operator on $mathbbF^n$. You've got a linear map between two unspecified spaces, so this formula will not apply. But, we can definitely use the same tools. Let
            beginalign*
            B_1 &= (e_1, e_2, e_3) \
            B_1' &= (e_1, e_1 + e_2, e_1 + e_2 + e_3) \
            B_2 &= (f_1, f_2) \
            B_2' &= (f_1, f_1 + f_2).
            endalign*

            We want $[L]_B_2' leftarrow B_1'$, and we know $[L]_B_2 leftarrow B_1$. We compute



            beginalign*
            [L]_B_2' leftarrow B_1' &= [mathrmid circ L circ mathrmid]_B_2' leftarrow B_1' \
            &= [mathrmid]_B_2' leftarrow B_2 [L]_B_2 leftarrow B_1 [mathrmid]_B_1 leftarrow B_1'.
            endalign*



            We know $[L]_B_2 leftarrow B_1$, so we must compute the other two matrices. We have



            $$[mathrmid]_B_1 leftarrow B_1' = left([e_1]_B_1 mid [e_1 + e_2]_B_1 | , [e_1 + e_2 + e_3]_B_1 right) = beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix.$$



            Similarly,



            $$[mathrmid]_B_2 leftarrow B_2' = left([f_1]_B_2 mid , [f_1 + f_2]_B_2right) = beginpmatrix 1 & 1 \ 0 & 1 endpmatrix,$$



            and so



            $$[mathrmid]_B_2' leftarrow B_2 = [mathrmid]^-1_B_2 leftarrow B_2' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix.$$



            Finally, this gives us,



            $$[L]_B_2' leftarrow B_1' = beginpmatrix 1 & -1 \ 0 & 1 endpmatrix beginpmatrix 0 & 1 & 2 \
            3 & 4 & 5 endpmatrix beginpmatrix 1 & 1 & 1 \ 0 & 1 & 1 \ 0 & 0 & 1 endpmatrix = beginpmatrix -3 & -6 & -9 \
            3 & 7 & 12 endpmatrix.$$







            share|cite|improve this answer












            share|cite|improve this answer



            share|cite|improve this answer










            answered 15 mins ago









            Theo Bendit

            14k12045




            14k12045











            • My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
              – Raj
              7 mins ago
















            • My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
              – Raj
              7 mins ago















            My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
            – Raj
            7 mins ago




            My thanks to you as well for the extremely in-depth answer, it's more than I can hope for. I think it will be especially helpful to see multiple deep perspectives on this.
            – Raj
            7 mins ago










            up vote
            1
            down vote













            You should have gotten
            $$
            D =
            left(
            beginarraycc
            1 & 1 \
            0 & 1
            endarray
            right)
            $$

            Then
            $$
            D^-1A C =
            left(
            beginarrayccc
            -3 & -6 & -9\
            3 & 7 & 12
            endarray
            right)
            $$






            share|cite|improve this answer
























              up vote
              1
              down vote













              You should have gotten
              $$
              D =
              left(
              beginarraycc
              1 & 1 \
              0 & 1
              endarray
              right)
              $$

              Then
              $$
              D^-1A C =
              left(
              beginarrayccc
              -3 & -6 & -9\
              3 & 7 & 12
              endarray
              right)
              $$






              share|cite|improve this answer






















                up vote
                1
                down vote










                up vote
                1
                down vote









                You should have gotten
                $$
                D =
                left(
                beginarraycc
                1 & 1 \
                0 & 1
                endarray
                right)
                $$

                Then
                $$
                D^-1A C =
                left(
                beginarrayccc
                -3 & -6 & -9\
                3 & 7 & 12
                endarray
                right)
                $$






                share|cite|improve this answer












                You should have gotten
                $$
                D =
                left(
                beginarraycc
                1 & 1 \
                0 & 1
                endarray
                right)
                $$

                Then
                $$
                D^-1A C =
                left(
                beginarrayccc
                -3 & -6 & -9\
                3 & 7 & 12
                endarray
                right)
                $$







                share|cite|improve this answer












                share|cite|improve this answer



                share|cite|improve this answer










                answered 1 hour ago









                Will Jagy

                98.4k595196




                98.4k595196



























                     

                    draft saved


                    draft discarded















































                     


                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2928327%2fchange-of-basis-of-a-linear-map-defined-by-non-square-matrix%23new-answer', 'question_page');

                    );

                    Post as a guest













































































                    Comments

                    Popular posts from this blog

                    Long meetings (6-7 hours a day): Being “babysat” by supervisor

                    Is the Concept of Multiple Fantasy Races Scientifically Flawed? [closed]

                    Confectionery