Elementary linear algebra and linear maps exercise.












1












$begingroup$


Let $E$ and $F$ be two finite dimensional vector spaces and $f: E to F$ a linear transformation. If $v_1, ... , v_r in E$ are linearly independent, prove that $operatorname*{Im} f = left langle f(v_1), ... , f(v_r) right rangle Leftrightarrow E = operatorname*{Ker} f + left langle v_1,... ,v_r right rangle$.



I have tried both using Grassman's formula and the rank-nullity theorem with the corresponding inequalities due to the fact that $v_1, ... , v_r in E$ are linearly independent, but I don't get a consistent proof for any of the implications.



NOTE: If $u_1, ... , u_n$ are vectors, $left langle u_1,...,u_n right rangle$ denotes the LINEAR SPAN of the vectors.










share|cite|improve this question











$endgroup$

















    1












    $begingroup$


    Let $E$ and $F$ be two finite dimensional vector spaces and $f: E to F$ a linear transformation. If $v_1, ... , v_r in E$ are linearly independent, prove that $operatorname*{Im} f = left langle f(v_1), ... , f(v_r) right rangle Leftrightarrow E = operatorname*{Ker} f + left langle v_1,... ,v_r right rangle$.



    I have tried both using Grassman's formula and the rank-nullity theorem with the corresponding inequalities due to the fact that $v_1, ... , v_r in E$ are linearly independent, but I don't get a consistent proof for any of the implications.



    NOTE: If $u_1, ... , u_n$ are vectors, $left langle u_1,...,u_n right rangle$ denotes the LINEAR SPAN of the vectors.










    share|cite|improve this question











    $endgroup$















      1












      1








      1


      1



      $begingroup$


      Let $E$ and $F$ be two finite dimensional vector spaces and $f: E to F$ a linear transformation. If $v_1, ... , v_r in E$ are linearly independent, prove that $operatorname*{Im} f = left langle f(v_1), ... , f(v_r) right rangle Leftrightarrow E = operatorname*{Ker} f + left langle v_1,... ,v_r right rangle$.



      I have tried both using Grassman's formula and the rank-nullity theorem with the corresponding inequalities due to the fact that $v_1, ... , v_r in E$ are linearly independent, but I don't get a consistent proof for any of the implications.



      NOTE: If $u_1, ... , u_n$ are vectors, $left langle u_1,...,u_n right rangle$ denotes the LINEAR SPAN of the vectors.










      share|cite|improve this question











      $endgroup$




      Let $E$ and $F$ be two finite dimensional vector spaces and $f: E to F$ a linear transformation. If $v_1, ... , v_r in E$ are linearly independent, prove that $operatorname*{Im} f = left langle f(v_1), ... , f(v_r) right rangle Leftrightarrow E = operatorname*{Ker} f + left langle v_1,... ,v_r right rangle$.



      I have tried both using Grassman's formula and the rank-nullity theorem with the corresponding inequalities due to the fact that $v_1, ... , v_r in E$ are linearly independent, but I don't get a consistent proof for any of the implications.



      NOTE: If $u_1, ... , u_n$ are vectors, $left langle u_1,...,u_n right rangle$ denotes the LINEAR SPAN of the vectors.







      linear-algebra linear-transformations






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited Jan 7 at 13:39







      Marc Ballestero Ribó

















      asked Jan 7 at 12:50









      Marc Ballestero RibóMarc Ballestero Ribó

      135




      135






















          2 Answers
          2






          active

          oldest

          votes


















          1












          $begingroup$


          I have tried both using Grassman's formula and the relationship between the dimension of the kernel, image, and domain of a linear map (dimE=dimImf+dimKerf) with the corresponding inequalities due to the fact that v1,...,vr∈E are linearly independent, but I don't get a consistent proof for any of the implications.




          None of those are going to be sufficient, because they're all results about dimensions, and we don't want a result about dimensions. There's a possibility that you could use it, combined with other results, but it's not actually needed here, and probably shouldn't be the first approach you think of.





          Instead, first note that, if $E = mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$, then we can choose some $w_1,ldots,w_{n-r}inmathrm{Ker}(f)$ to extend $langle v_1,ldots,v_rrangle$ to a basis of $E$. Then for any $u = sum alpha_iv_i + sumbeta_iw_i$, we have $f(u) = sumalpha_if(v_i)+sumbeta_if(w_i) = sumalpha_if(v_i)$ since $w_iinmathrm{Ker}(f)$, and note that this lies in $langle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)subseteqlangle f(v_1),ldots,f(v_r)rangle$. The reverse inclusion is simple.



          For the reverse implication, suppose that $E neq mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Then there is some $u in E$ that does not lie in $mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Now, if $f(u)$ lies in $langle f(v_1),ldots,f(v_r)rangle$, then there are some $alpha_1,ldots,alpha_r$ such that $f(u) = sumalpha_if(v_i) = f(sumalpha_iv_i)$. Thus, $w := u - sumalpha_iv_i$ has $f(w) = f(u) - f(sumalpha_iv_i) = 0$, so $winmathrm{Ker}(f)$. But then, $u = w + sumalpha_iv_i$, so $uinmathrm{Ker}(f)+langle v_1,ldots,v_rrangle$, a contradiction. Thus, $f(u)notinlangle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)neq langle f(v_1),ldots,f(v_r)rangle$.






          share|cite|improve this answer









          $endgroup$





















            1












            $begingroup$

            Let $U=langle v_1,dots,v_rrangle$ and $K=ker f$.



            Grassmann’s formula tells you that $dim U+dim K=dim(U+K)+dim(Ucap K)$, so
            $$
            dim(U+K)=dim U+dim K-dim(Ucap K)=r+n-k
            $$

            where $n=dim K$ and $k=dim(Kcap N)$; let $s$ be the rank of $f$, so $dim E=s+n$ by the rank-nullity theorem.



            We have $E=U+K$ if and only if $r+n-k=s+n$, that is, if and only if $r=k+s$.



            We can also consider $f'colon Uto F$ (the restriction of $f$). The nullity of $f'$ is $k$; if $s'$ is the rank of $f'$, then $r=k+s'$. Clearly the image of $f'$ is a subspace of the image of $f$ and is spanned by $f(v_1),dots,f(v_r)$.



            Therefore $r=k+s$ if and only if $s=s'$ that is, if and only if $operatorname{im}f=operatorname{im}f'=langle f(v_1),dots,f(v_r)rangle$.






            share|cite|improve this answer









            $endgroup$













              Your Answer





              StackExchange.ifUsing("editor", function () {
              return StackExchange.using("mathjaxEditing", function () {
              StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
              StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
              });
              });
              }, "mathjax-editing");

              StackExchange.ready(function() {
              var channelOptions = {
              tags: "".split(" "),
              id: "69"
              };
              initTagRenderer("".split(" "), "".split(" "), channelOptions);

              StackExchange.using("externalEditor", function() {
              // Have to fire editor after snippets, if snippets enabled
              if (StackExchange.settings.snippets.snippetsEnabled) {
              StackExchange.using("snippets", function() {
              createEditor();
              });
              }
              else {
              createEditor();
              }
              });

              function createEditor() {
              StackExchange.prepareEditor({
              heartbeatType: 'answer',
              autoActivateHeartbeat: false,
              convertImagesToLinks: true,
              noModals: true,
              showLowRepImageUploadWarning: true,
              reputationToPostImages: 10,
              bindNavPrevention: true,
              postfix: "",
              imageUploader: {
              brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
              contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
              allowUrls: true
              },
              noCode: true, onDemand: true,
              discardSelector: ".discard-answer"
              ,immediatelyShowMarkdownHelp:true
              });


              }
              });














              draft saved

              draft discarded


















              StackExchange.ready(
              function () {
              StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3064972%2felementary-linear-algebra-and-linear-maps-exercise%23new-answer', 'question_page');
              }
              );

              Post as a guest















              Required, but never shown

























              2 Answers
              2






              active

              oldest

              votes








              2 Answers
              2






              active

              oldest

              votes









              active

              oldest

              votes






              active

              oldest

              votes









              1












              $begingroup$


              I have tried both using Grassman's formula and the relationship between the dimension of the kernel, image, and domain of a linear map (dimE=dimImf+dimKerf) with the corresponding inequalities due to the fact that v1,...,vr∈E are linearly independent, but I don't get a consistent proof for any of the implications.




              None of those are going to be sufficient, because they're all results about dimensions, and we don't want a result about dimensions. There's a possibility that you could use it, combined with other results, but it's not actually needed here, and probably shouldn't be the first approach you think of.





              Instead, first note that, if $E = mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$, then we can choose some $w_1,ldots,w_{n-r}inmathrm{Ker}(f)$ to extend $langle v_1,ldots,v_rrangle$ to a basis of $E$. Then for any $u = sum alpha_iv_i + sumbeta_iw_i$, we have $f(u) = sumalpha_if(v_i)+sumbeta_if(w_i) = sumalpha_if(v_i)$ since $w_iinmathrm{Ker}(f)$, and note that this lies in $langle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)subseteqlangle f(v_1),ldots,f(v_r)rangle$. The reverse inclusion is simple.



              For the reverse implication, suppose that $E neq mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Then there is some $u in E$ that does not lie in $mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Now, if $f(u)$ lies in $langle f(v_1),ldots,f(v_r)rangle$, then there are some $alpha_1,ldots,alpha_r$ such that $f(u) = sumalpha_if(v_i) = f(sumalpha_iv_i)$. Thus, $w := u - sumalpha_iv_i$ has $f(w) = f(u) - f(sumalpha_iv_i) = 0$, so $winmathrm{Ker}(f)$. But then, $u = w + sumalpha_iv_i$, so $uinmathrm{Ker}(f)+langle v_1,ldots,v_rrangle$, a contradiction. Thus, $f(u)notinlangle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)neq langle f(v_1),ldots,f(v_r)rangle$.






              share|cite|improve this answer









              $endgroup$


















                1












                $begingroup$


                I have tried both using Grassman's formula and the relationship between the dimension of the kernel, image, and domain of a linear map (dimE=dimImf+dimKerf) with the corresponding inequalities due to the fact that v1,...,vr∈E are linearly independent, but I don't get a consistent proof for any of the implications.




                None of those are going to be sufficient, because they're all results about dimensions, and we don't want a result about dimensions. There's a possibility that you could use it, combined with other results, but it's not actually needed here, and probably shouldn't be the first approach you think of.





                Instead, first note that, if $E = mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$, then we can choose some $w_1,ldots,w_{n-r}inmathrm{Ker}(f)$ to extend $langle v_1,ldots,v_rrangle$ to a basis of $E$. Then for any $u = sum alpha_iv_i + sumbeta_iw_i$, we have $f(u) = sumalpha_if(v_i)+sumbeta_if(w_i) = sumalpha_if(v_i)$ since $w_iinmathrm{Ker}(f)$, and note that this lies in $langle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)subseteqlangle f(v_1),ldots,f(v_r)rangle$. The reverse inclusion is simple.



                For the reverse implication, suppose that $E neq mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Then there is some $u in E$ that does not lie in $mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Now, if $f(u)$ lies in $langle f(v_1),ldots,f(v_r)rangle$, then there are some $alpha_1,ldots,alpha_r$ such that $f(u) = sumalpha_if(v_i) = f(sumalpha_iv_i)$. Thus, $w := u - sumalpha_iv_i$ has $f(w) = f(u) - f(sumalpha_iv_i) = 0$, so $winmathrm{Ker}(f)$. But then, $u = w + sumalpha_iv_i$, so $uinmathrm{Ker}(f)+langle v_1,ldots,v_rrangle$, a contradiction. Thus, $f(u)notinlangle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)neq langle f(v_1),ldots,f(v_r)rangle$.






                share|cite|improve this answer









                $endgroup$
















                  1












                  1








                  1





                  $begingroup$


                  I have tried both using Grassman's formula and the relationship between the dimension of the kernel, image, and domain of a linear map (dimE=dimImf+dimKerf) with the corresponding inequalities due to the fact that v1,...,vr∈E are linearly independent, but I don't get a consistent proof for any of the implications.




                  None of those are going to be sufficient, because they're all results about dimensions, and we don't want a result about dimensions. There's a possibility that you could use it, combined with other results, but it's not actually needed here, and probably shouldn't be the first approach you think of.





                  Instead, first note that, if $E = mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$, then we can choose some $w_1,ldots,w_{n-r}inmathrm{Ker}(f)$ to extend $langle v_1,ldots,v_rrangle$ to a basis of $E$. Then for any $u = sum alpha_iv_i + sumbeta_iw_i$, we have $f(u) = sumalpha_if(v_i)+sumbeta_if(w_i) = sumalpha_if(v_i)$ since $w_iinmathrm{Ker}(f)$, and note that this lies in $langle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)subseteqlangle f(v_1),ldots,f(v_r)rangle$. The reverse inclusion is simple.



                  For the reverse implication, suppose that $E neq mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Then there is some $u in E$ that does not lie in $mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Now, if $f(u)$ lies in $langle f(v_1),ldots,f(v_r)rangle$, then there are some $alpha_1,ldots,alpha_r$ such that $f(u) = sumalpha_if(v_i) = f(sumalpha_iv_i)$. Thus, $w := u - sumalpha_iv_i$ has $f(w) = f(u) - f(sumalpha_iv_i) = 0$, so $winmathrm{Ker}(f)$. But then, $u = w + sumalpha_iv_i$, so $uinmathrm{Ker}(f)+langle v_1,ldots,v_rrangle$, a contradiction. Thus, $f(u)notinlangle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)neq langle f(v_1),ldots,f(v_r)rangle$.






                  share|cite|improve this answer









                  $endgroup$




                  I have tried both using Grassman's formula and the relationship between the dimension of the kernel, image, and domain of a linear map (dimE=dimImf+dimKerf) with the corresponding inequalities due to the fact that v1,...,vr∈E are linearly independent, but I don't get a consistent proof for any of the implications.




                  None of those are going to be sufficient, because they're all results about dimensions, and we don't want a result about dimensions. There's a possibility that you could use it, combined with other results, but it's not actually needed here, and probably shouldn't be the first approach you think of.





                  Instead, first note that, if $E = mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$, then we can choose some $w_1,ldots,w_{n-r}inmathrm{Ker}(f)$ to extend $langle v_1,ldots,v_rrangle$ to a basis of $E$. Then for any $u = sum alpha_iv_i + sumbeta_iw_i$, we have $f(u) = sumalpha_if(v_i)+sumbeta_if(w_i) = sumalpha_if(v_i)$ since $w_iinmathrm{Ker}(f)$, and note that this lies in $langle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)subseteqlangle f(v_1),ldots,f(v_r)rangle$. The reverse inclusion is simple.



                  For the reverse implication, suppose that $E neq mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Then there is some $u in E$ that does not lie in $mathrm{Ker}(f) + langle v_1,ldots,v_rrangle$. Now, if $f(u)$ lies in $langle f(v_1),ldots,f(v_r)rangle$, then there are some $alpha_1,ldots,alpha_r$ such that $f(u) = sumalpha_if(v_i) = f(sumalpha_iv_i)$. Thus, $w := u - sumalpha_iv_i$ has $f(w) = f(u) - f(sumalpha_iv_i) = 0$, so $winmathrm{Ker}(f)$. But then, $u = w + sumalpha_iv_i$, so $uinmathrm{Ker}(f)+langle v_1,ldots,v_rrangle$, a contradiction. Thus, $f(u)notinlangle f(v_1),ldots,f(v_r)rangle$, so $mathrm{Im}(f)neq langle f(v_1),ldots,f(v_r)rangle$.







                  share|cite|improve this answer












                  share|cite|improve this answer



                  share|cite|improve this answer










                  answered Jan 7 at 13:28









                  user3482749user3482749

                  3,554417




                  3,554417























                      1












                      $begingroup$

                      Let $U=langle v_1,dots,v_rrangle$ and $K=ker f$.



                      Grassmann’s formula tells you that $dim U+dim K=dim(U+K)+dim(Ucap K)$, so
                      $$
                      dim(U+K)=dim U+dim K-dim(Ucap K)=r+n-k
                      $$

                      where $n=dim K$ and $k=dim(Kcap N)$; let $s$ be the rank of $f$, so $dim E=s+n$ by the rank-nullity theorem.



                      We have $E=U+K$ if and only if $r+n-k=s+n$, that is, if and only if $r=k+s$.



                      We can also consider $f'colon Uto F$ (the restriction of $f$). The nullity of $f'$ is $k$; if $s'$ is the rank of $f'$, then $r=k+s'$. Clearly the image of $f'$ is a subspace of the image of $f$ and is spanned by $f(v_1),dots,f(v_r)$.



                      Therefore $r=k+s$ if and only if $s=s'$ that is, if and only if $operatorname{im}f=operatorname{im}f'=langle f(v_1),dots,f(v_r)rangle$.






                      share|cite|improve this answer









                      $endgroup$


















                        1












                        $begingroup$

                        Let $U=langle v_1,dots,v_rrangle$ and $K=ker f$.



                        Grassmann’s formula tells you that $dim U+dim K=dim(U+K)+dim(Ucap K)$, so
                        $$
                        dim(U+K)=dim U+dim K-dim(Ucap K)=r+n-k
                        $$

                        where $n=dim K$ and $k=dim(Kcap N)$; let $s$ be the rank of $f$, so $dim E=s+n$ by the rank-nullity theorem.



                        We have $E=U+K$ if and only if $r+n-k=s+n$, that is, if and only if $r=k+s$.



                        We can also consider $f'colon Uto F$ (the restriction of $f$). The nullity of $f'$ is $k$; if $s'$ is the rank of $f'$, then $r=k+s'$. Clearly the image of $f'$ is a subspace of the image of $f$ and is spanned by $f(v_1),dots,f(v_r)$.



                        Therefore $r=k+s$ if and only if $s=s'$ that is, if and only if $operatorname{im}f=operatorname{im}f'=langle f(v_1),dots,f(v_r)rangle$.






                        share|cite|improve this answer









                        $endgroup$
















                          1












                          1








                          1





                          $begingroup$

                          Let $U=langle v_1,dots,v_rrangle$ and $K=ker f$.



                          Grassmann’s formula tells you that $dim U+dim K=dim(U+K)+dim(Ucap K)$, so
                          $$
                          dim(U+K)=dim U+dim K-dim(Ucap K)=r+n-k
                          $$

                          where $n=dim K$ and $k=dim(Kcap N)$; let $s$ be the rank of $f$, so $dim E=s+n$ by the rank-nullity theorem.



                          We have $E=U+K$ if and only if $r+n-k=s+n$, that is, if and only if $r=k+s$.



                          We can also consider $f'colon Uto F$ (the restriction of $f$). The nullity of $f'$ is $k$; if $s'$ is the rank of $f'$, then $r=k+s'$. Clearly the image of $f'$ is a subspace of the image of $f$ and is spanned by $f(v_1),dots,f(v_r)$.



                          Therefore $r=k+s$ if and only if $s=s'$ that is, if and only if $operatorname{im}f=operatorname{im}f'=langle f(v_1),dots,f(v_r)rangle$.






                          share|cite|improve this answer









                          $endgroup$



                          Let $U=langle v_1,dots,v_rrangle$ and $K=ker f$.



                          Grassmann’s formula tells you that $dim U+dim K=dim(U+K)+dim(Ucap K)$, so
                          $$
                          dim(U+K)=dim U+dim K-dim(Ucap K)=r+n-k
                          $$

                          where $n=dim K$ and $k=dim(Kcap N)$; let $s$ be the rank of $f$, so $dim E=s+n$ by the rank-nullity theorem.



                          We have $E=U+K$ if and only if $r+n-k=s+n$, that is, if and only if $r=k+s$.



                          We can also consider $f'colon Uto F$ (the restriction of $f$). The nullity of $f'$ is $k$; if $s'$ is the rank of $f'$, then $r=k+s'$. Clearly the image of $f'$ is a subspace of the image of $f$ and is spanned by $f(v_1),dots,f(v_r)$.



                          Therefore $r=k+s$ if and only if $s=s'$ that is, if and only if $operatorname{im}f=operatorname{im}f'=langle f(v_1),dots,f(v_r)rangle$.







                          share|cite|improve this answer












                          share|cite|improve this answer



                          share|cite|improve this answer










                          answered Jan 7 at 15:17









                          egregegreg

                          180k1485202




                          180k1485202






























                              draft saved

                              draft discarded




















































                              Thanks for contributing an answer to Mathematics Stack Exchange!


                              • Please be sure to answer the question. Provide details and share your research!

                              But avoid



                              • Asking for help, clarification, or responding to other answers.

                              • Making statements based on opinion; back them up with references or personal experience.


                              Use MathJax to format equations. MathJax reference.


                              To learn more, see our tips on writing great answers.




                              draft saved


                              draft discarded














                              StackExchange.ready(
                              function () {
                              StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3064972%2felementary-linear-algebra-and-linear-maps-exercise%23new-answer', 'question_page');
                              }
                              );

                              Post as a guest















                              Required, but never shown





















































                              Required, but never shown














                              Required, but never shown












                              Required, but never shown







                              Required, but never shown

































                              Required, but never shown














                              Required, but never shown












                              Required, but never shown







                              Required, but never shown







                              Popular posts from this blog

                              Mario Kart Wii

                              What does “Dominus providebit” mean?

                              Antonio Litta Visconti Arese