What inspires people to define linear maps?












4












$begingroup$


I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.



My understanding now is that a map is like a function that maps something to another thing.



Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?



Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.



And finally, why linear maps have the additivity and homogeneity properties?










share|cite|improve this question









$endgroup$

















    4












    $begingroup$


    I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.



    My understanding now is that a map is like a function that maps something to another thing.



    Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?



    Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.



    And finally, why linear maps have the additivity and homogeneity properties?










    share|cite|improve this question









    $endgroup$















      4












      4








      4





      $begingroup$


      I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.



      My understanding now is that a map is like a function that maps something to another thing.



      Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?



      Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.



      And finally, why linear maps have the additivity and homogeneity properties?










      share|cite|improve this question









      $endgroup$




      I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.



      My understanding now is that a map is like a function that maps something to another thing.



      Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?



      Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.



      And finally, why linear maps have the additivity and homogeneity properties?







      linear-algebra






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked 5 hours ago









      JOHN JOHN

      1207




      1207






















          2 Answers
          2






          active

          oldest

          votes


















          2












          $begingroup$

          Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



          The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



          In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



          Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



          Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



          In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



          Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



          I hope that answers your question.






          share|cite|improve this answer









          $endgroup$





















            4












            $begingroup$

            In calculus, the derivative is defined by
            $$
            f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
            $$

            Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
            $$
            tag{1} Delta f approx f'(x_0) Delta x.
            $$



            We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
            $$
            underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
            $$

            The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



            In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



            The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
            $$
            f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
            $$



            When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



            With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.






            share|cite|improve this answer











            $endgroup$













              Your Answer





              StackExchange.ifUsing("editor", function () {
              return StackExchange.using("mathjaxEditing", function () {
              StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
              StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
              });
              });
              }, "mathjax-editing");

              StackExchange.ready(function() {
              var channelOptions = {
              tags: "".split(" "),
              id: "69"
              };
              initTagRenderer("".split(" "), "".split(" "), channelOptions);

              StackExchange.using("externalEditor", function() {
              // Have to fire editor after snippets, if snippets enabled
              if (StackExchange.settings.snippets.snippetsEnabled) {
              StackExchange.using("snippets", function() {
              createEditor();
              });
              }
              else {
              createEditor();
              }
              });

              function createEditor() {
              StackExchange.prepareEditor({
              heartbeatType: 'answer',
              autoActivateHeartbeat: false,
              convertImagesToLinks: true,
              noModals: true,
              showLowRepImageUploadWarning: true,
              reputationToPostImages: 10,
              bindNavPrevention: true,
              postfix: "",
              imageUploader: {
              brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
              contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
              allowUrls: true
              },
              noCode: true, onDemand: true,
              discardSelector: ".discard-answer"
              ,immediatelyShowMarkdownHelp:true
              });


              }
              });














              draft saved

              draft discarded


















              StackExchange.ready(
              function () {
              StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3070526%2fwhat-inspires-people-to-define-linear-maps%23new-answer', 'question_page');
              }
              );

              Post as a guest















              Required, but never shown

























              2 Answers
              2






              active

              oldest

              votes








              2 Answers
              2






              active

              oldest

              votes









              active

              oldest

              votes






              active

              oldest

              votes









              2












              $begingroup$

              Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



              The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



              In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



              Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



              Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



              In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



              Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



              I hope that answers your question.






              share|cite|improve this answer









              $endgroup$


















                2












                $begingroup$

                Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



                The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



                In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



                Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



                Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



                In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



                Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



                I hope that answers your question.






                share|cite|improve this answer









                $endgroup$
















                  2












                  2








                  2





                  $begingroup$

                  Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



                  The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



                  In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



                  Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



                  Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



                  In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



                  Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



                  I hope that answers your question.






                  share|cite|improve this answer









                  $endgroup$



                  Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



                  The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



                  In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



                  Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



                  Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



                  In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



                  Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



                  I hope that answers your question.







                  share|cite|improve this answer












                  share|cite|improve this answer



                  share|cite|improve this answer










                  answered 4 hours ago









                  Theo BenditTheo Bendit

                  17k12148




                  17k12148























                      4












                      $begingroup$

                      In calculus, the derivative is defined by
                      $$
                      f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
                      $$

                      Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
                      $$
                      tag{1} Delta f approx f'(x_0) Delta x.
                      $$



                      We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
                      $$
                      underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
                      $$

                      The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



                      In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



                      The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
                      $$
                      f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
                      $$



                      When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



                      With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.






                      share|cite|improve this answer











                      $endgroup$


















                        4












                        $begingroup$

                        In calculus, the derivative is defined by
                        $$
                        f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
                        $$

                        Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
                        $$
                        tag{1} Delta f approx f'(x_0) Delta x.
                        $$



                        We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
                        $$
                        underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
                        $$

                        The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



                        In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



                        The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
                        $$
                        f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
                        $$



                        When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



                        With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.






                        share|cite|improve this answer











                        $endgroup$
















                          4












                          4








                          4





                          $begingroup$

                          In calculus, the derivative is defined by
                          $$
                          f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
                          $$

                          Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
                          $$
                          tag{1} Delta f approx f'(x_0) Delta x.
                          $$



                          We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
                          $$
                          underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
                          $$

                          The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



                          In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



                          The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
                          $$
                          f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
                          $$



                          When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



                          With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.






                          share|cite|improve this answer











                          $endgroup$



                          In calculus, the derivative is defined by
                          $$
                          f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
                          $$

                          Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
                          $$
                          tag{1} Delta f approx f'(x_0) Delta x.
                          $$



                          We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
                          $$
                          underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
                          $$

                          The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



                          In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



                          The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
                          $$
                          f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
                          $$



                          When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



                          With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.







                          share|cite|improve this answer














                          share|cite|improve this answer



                          share|cite|improve this answer








                          edited 4 hours ago

























                          answered 4 hours ago









                          littleOlittleO

                          29.3k644109




                          29.3k644109






























                              draft saved

                              draft discarded




















































                              Thanks for contributing an answer to Mathematics Stack Exchange!


                              • Please be sure to answer the question. Provide details and share your research!

                              But avoid



                              • Asking for help, clarification, or responding to other answers.

                              • Making statements based on opinion; back them up with references or personal experience.


                              Use MathJax to format equations. MathJax reference.


                              To learn more, see our tips on writing great answers.




                              draft saved


                              draft discarded














                              StackExchange.ready(
                              function () {
                              StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3070526%2fwhat-inspires-people-to-define-linear-maps%23new-answer', 'question_page');
                              }
                              );

                              Post as a guest















                              Required, but never shown





















































                              Required, but never shown














                              Required, but never shown












                              Required, but never shown







                              Required, but never shown

































                              Required, but never shown














                              Required, but never shown












                              Required, but never shown







                              Required, but never shown







                              Popular posts from this blog

                              Plaza Victoria

                              Puebla de Zaragoza

                              Musa