Order of convergence of a two-step iteration method












0












$begingroup$


Find the order of the iteration is given by



$$y_{n+1} = x_n - frac{f(x_n)}{f'(x_n)} quad x_{n+1} = y_{n+1} - frac{f(y_{n+1})}{f'(x_n)}$$



Assuming $f(r) = 0$, $f'(r) neq 0$ and the initial guess is close to r.



I know the iteration is a third order method because if I let $Y(x) = x - frac{f(x)}{f'(x)}$ and $I(x) = Y(x) - frac{f(Y(x))}{f'(x)}$. We end up with
$$frac{dI(x)}{dx}|_r = 0$$
$$frac{d^2I(x)}{dx^2}|_r = 0$$
$$frac{d^3I(x)}{dx^3}|_r neq 0$$



This way however is very messy when evaluating all of the derivatives. I was wondering if there is another way to prove that the above methods order of convergence?










share|cite|improve this question











$endgroup$

















    0












    $begingroup$


    Find the order of the iteration is given by



    $$y_{n+1} = x_n - frac{f(x_n)}{f'(x_n)} quad x_{n+1} = y_{n+1} - frac{f(y_{n+1})}{f'(x_n)}$$



    Assuming $f(r) = 0$, $f'(r) neq 0$ and the initial guess is close to r.



    I know the iteration is a third order method because if I let $Y(x) = x - frac{f(x)}{f'(x)}$ and $I(x) = Y(x) - frac{f(Y(x))}{f'(x)}$. We end up with
    $$frac{dI(x)}{dx}|_r = 0$$
    $$frac{d^2I(x)}{dx^2}|_r = 0$$
    $$frac{d^3I(x)}{dx^3}|_r neq 0$$



    This way however is very messy when evaluating all of the derivatives. I was wondering if there is another way to prove that the above methods order of convergence?










    share|cite|improve this question











    $endgroup$















      0












      0








      0





      $begingroup$


      Find the order of the iteration is given by



      $$y_{n+1} = x_n - frac{f(x_n)}{f'(x_n)} quad x_{n+1} = y_{n+1} - frac{f(y_{n+1})}{f'(x_n)}$$



      Assuming $f(r) = 0$, $f'(r) neq 0$ and the initial guess is close to r.



      I know the iteration is a third order method because if I let $Y(x) = x - frac{f(x)}{f'(x)}$ and $I(x) = Y(x) - frac{f(Y(x))}{f'(x)}$. We end up with
      $$frac{dI(x)}{dx}|_r = 0$$
      $$frac{d^2I(x)}{dx^2}|_r = 0$$
      $$frac{d^3I(x)}{dx^3}|_r neq 0$$



      This way however is very messy when evaluating all of the derivatives. I was wondering if there is another way to prove that the above methods order of convergence?










      share|cite|improve this question











      $endgroup$




      Find the order of the iteration is given by



      $$y_{n+1} = x_n - frac{f(x_n)}{f'(x_n)} quad x_{n+1} = y_{n+1} - frac{f(y_{n+1})}{f'(x_n)}$$



      Assuming $f(r) = 0$, $f'(r) neq 0$ and the initial guess is close to r.



      I know the iteration is a third order method because if I let $Y(x) = x - frac{f(x)}{f'(x)}$ and $I(x) = Y(x) - frac{f(Y(x))}{f'(x)}$. We end up with
      $$frac{dI(x)}{dx}|_r = 0$$
      $$frac{d^2I(x)}{dx^2}|_r = 0$$
      $$frac{d^3I(x)}{dx^3}|_r neq 0$$



      This way however is very messy when evaluating all of the derivatives. I was wondering if there is another way to prove that the above methods order of convergence?







      numerical-methods






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited Dec 16 '18 at 19:44









      Rebellos

      15.3k31250




      15.3k31250










      asked Dec 16 '18 at 19:40







      user627004





























          1 Answer
          1






          active

          oldest

          votes


















          1












          $begingroup$

          Let $L$ be a Lipschitz constant of $f'$ on the domain under consideration. Then we know that
          $$
          |f(x+v)-f(x)-f'(x)v|leint_0^1|f'(x+sv)-f'(x)|,|v|,dsle L,|v|,int_0^1|sv|,ds=frac{L}2|v|^2,
          $$

          so that especially
          $$
          |f(y_{n+1})|lefrac{L}2frac{|f(x_n)|^2}{|f'(x_n)|^2}
          $$

          Further,
          $$
          f(x_{n+1})=f(y_{n+1})-f'(y_{n+1})frac{f(y_{n+1})}{f'(x_n)}+R~~text{with}~~ |R|lefrac{L}2frac{|f(y_{n+1}|)^2}{|f'(x_n)|^2}
          $$

          so that
          $$
          |f(x_{n+1})|
          le L,|x_{n+1}-y_{n+1}|,frac{|f(y_{n+1})|}{|f'(x_n)|}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
          le frac{L^2}{2}frac{|f(x_n)|^3}{|f'(x_n)|^4}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
          $$

          As the function value stands as proxy for the distance to the root, this shows the third order convergence.





          Note however that in terms of effort, or the Ostrowski index, you only get for polynomial $f$ an order of $sqrt[3]{3}=1.442$ per function and derivative evaluation, in the non-polynomial case even only $sqrt[4]3=1.316$ as one derivative costs as much as 2 function evaluations. To compare, the Newton method has $sqrt2=1.414$ resp. $sqrt[3]2=1.260$ and the secant method $frac{1+sqrt5}2=1.618$ (if it converges).



          So this method is slightly better than the Newton method. To put it in integer terms, in the polynomial case within the effort of $6$ function evaluations one can perform




          • 3 Newton steps with error reduction from $epsilon$ to $ϵ^8$ or

          • 2 cycles of this two-step method with an error reduction to $ϵ^9$.


          In the non-polynomial case, in the time of $12$ evaluations of $f$ one can perform





          • $4$ Newton steps to $ϵ^{16}$ or


          • $3$ two-step cycles to $ϵ^{27}$.






          share|cite|improve this answer











          $endgroup$













            Your Answer





            StackExchange.ifUsing("editor", function () {
            return StackExchange.using("mathjaxEditing", function () {
            StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
            StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
            });
            });
            }, "mathjax-editing");

            StackExchange.ready(function() {
            var channelOptions = {
            tags: "".split(" "),
            id: "69"
            };
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function() {
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled) {
            StackExchange.using("snippets", function() {
            createEditor();
            });
            }
            else {
            createEditor();
            }
            });

            function createEditor() {
            StackExchange.prepareEditor({
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader: {
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            },
            noCode: true, onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            });


            }
            });














            draft saved

            draft discarded


















            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3043066%2forder-of-convergence-of-a-two-step-iteration-method%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown
























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            1












            $begingroup$

            Let $L$ be a Lipschitz constant of $f'$ on the domain under consideration. Then we know that
            $$
            |f(x+v)-f(x)-f'(x)v|leint_0^1|f'(x+sv)-f'(x)|,|v|,dsle L,|v|,int_0^1|sv|,ds=frac{L}2|v|^2,
            $$

            so that especially
            $$
            |f(y_{n+1})|lefrac{L}2frac{|f(x_n)|^2}{|f'(x_n)|^2}
            $$

            Further,
            $$
            f(x_{n+1})=f(y_{n+1})-f'(y_{n+1})frac{f(y_{n+1})}{f'(x_n)}+R~~text{with}~~ |R|lefrac{L}2frac{|f(y_{n+1}|)^2}{|f'(x_n)|^2}
            $$

            so that
            $$
            |f(x_{n+1})|
            le L,|x_{n+1}-y_{n+1}|,frac{|f(y_{n+1})|}{|f'(x_n)|}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
            le frac{L^2}{2}frac{|f(x_n)|^3}{|f'(x_n)|^4}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
            $$

            As the function value stands as proxy for the distance to the root, this shows the third order convergence.





            Note however that in terms of effort, or the Ostrowski index, you only get for polynomial $f$ an order of $sqrt[3]{3}=1.442$ per function and derivative evaluation, in the non-polynomial case even only $sqrt[4]3=1.316$ as one derivative costs as much as 2 function evaluations. To compare, the Newton method has $sqrt2=1.414$ resp. $sqrt[3]2=1.260$ and the secant method $frac{1+sqrt5}2=1.618$ (if it converges).



            So this method is slightly better than the Newton method. To put it in integer terms, in the polynomial case within the effort of $6$ function evaluations one can perform




            • 3 Newton steps with error reduction from $epsilon$ to $ϵ^8$ or

            • 2 cycles of this two-step method with an error reduction to $ϵ^9$.


            In the non-polynomial case, in the time of $12$ evaluations of $f$ one can perform





            • $4$ Newton steps to $ϵ^{16}$ or


            • $3$ two-step cycles to $ϵ^{27}$.






            share|cite|improve this answer











            $endgroup$


















              1












              $begingroup$

              Let $L$ be a Lipschitz constant of $f'$ on the domain under consideration. Then we know that
              $$
              |f(x+v)-f(x)-f'(x)v|leint_0^1|f'(x+sv)-f'(x)|,|v|,dsle L,|v|,int_0^1|sv|,ds=frac{L}2|v|^2,
              $$

              so that especially
              $$
              |f(y_{n+1})|lefrac{L}2frac{|f(x_n)|^2}{|f'(x_n)|^2}
              $$

              Further,
              $$
              f(x_{n+1})=f(y_{n+1})-f'(y_{n+1})frac{f(y_{n+1})}{f'(x_n)}+R~~text{with}~~ |R|lefrac{L}2frac{|f(y_{n+1}|)^2}{|f'(x_n)|^2}
              $$

              so that
              $$
              |f(x_{n+1})|
              le L,|x_{n+1}-y_{n+1}|,frac{|f(y_{n+1})|}{|f'(x_n)|}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
              le frac{L^2}{2}frac{|f(x_n)|^3}{|f'(x_n)|^4}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
              $$

              As the function value stands as proxy for the distance to the root, this shows the third order convergence.





              Note however that in terms of effort, or the Ostrowski index, you only get for polynomial $f$ an order of $sqrt[3]{3}=1.442$ per function and derivative evaluation, in the non-polynomial case even only $sqrt[4]3=1.316$ as one derivative costs as much as 2 function evaluations. To compare, the Newton method has $sqrt2=1.414$ resp. $sqrt[3]2=1.260$ and the secant method $frac{1+sqrt5}2=1.618$ (if it converges).



              So this method is slightly better than the Newton method. To put it in integer terms, in the polynomial case within the effort of $6$ function evaluations one can perform




              • 3 Newton steps with error reduction from $epsilon$ to $ϵ^8$ or

              • 2 cycles of this two-step method with an error reduction to $ϵ^9$.


              In the non-polynomial case, in the time of $12$ evaluations of $f$ one can perform





              • $4$ Newton steps to $ϵ^{16}$ or


              • $3$ two-step cycles to $ϵ^{27}$.






              share|cite|improve this answer











              $endgroup$
















                1












                1








                1





                $begingroup$

                Let $L$ be a Lipschitz constant of $f'$ on the domain under consideration. Then we know that
                $$
                |f(x+v)-f(x)-f'(x)v|leint_0^1|f'(x+sv)-f'(x)|,|v|,dsle L,|v|,int_0^1|sv|,ds=frac{L}2|v|^2,
                $$

                so that especially
                $$
                |f(y_{n+1})|lefrac{L}2frac{|f(x_n)|^2}{|f'(x_n)|^2}
                $$

                Further,
                $$
                f(x_{n+1})=f(y_{n+1})-f'(y_{n+1})frac{f(y_{n+1})}{f'(x_n)}+R~~text{with}~~ |R|lefrac{L}2frac{|f(y_{n+1}|)^2}{|f'(x_n)|^2}
                $$

                so that
                $$
                |f(x_{n+1})|
                le L,|x_{n+1}-y_{n+1}|,frac{|f(y_{n+1})|}{|f'(x_n)|}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
                le frac{L^2}{2}frac{|f(x_n)|^3}{|f'(x_n)|^4}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
                $$

                As the function value stands as proxy for the distance to the root, this shows the third order convergence.





                Note however that in terms of effort, or the Ostrowski index, you only get for polynomial $f$ an order of $sqrt[3]{3}=1.442$ per function and derivative evaluation, in the non-polynomial case even only $sqrt[4]3=1.316$ as one derivative costs as much as 2 function evaluations. To compare, the Newton method has $sqrt2=1.414$ resp. $sqrt[3]2=1.260$ and the secant method $frac{1+sqrt5}2=1.618$ (if it converges).



                So this method is slightly better than the Newton method. To put it in integer terms, in the polynomial case within the effort of $6$ function evaluations one can perform




                • 3 Newton steps with error reduction from $epsilon$ to $ϵ^8$ or

                • 2 cycles of this two-step method with an error reduction to $ϵ^9$.


                In the non-polynomial case, in the time of $12$ evaluations of $f$ one can perform





                • $4$ Newton steps to $ϵ^{16}$ or


                • $3$ two-step cycles to $ϵ^{27}$.






                share|cite|improve this answer











                $endgroup$



                Let $L$ be a Lipschitz constant of $f'$ on the domain under consideration. Then we know that
                $$
                |f(x+v)-f(x)-f'(x)v|leint_0^1|f'(x+sv)-f'(x)|,|v|,dsle L,|v|,int_0^1|sv|,ds=frac{L}2|v|^2,
                $$

                so that especially
                $$
                |f(y_{n+1})|lefrac{L}2frac{|f(x_n)|^2}{|f'(x_n)|^2}
                $$

                Further,
                $$
                f(x_{n+1})=f(y_{n+1})-f'(y_{n+1})frac{f(y_{n+1})}{f'(x_n)}+R~~text{with}~~ |R|lefrac{L}2frac{|f(y_{n+1}|)^2}{|f'(x_n)|^2}
                $$

                so that
                $$
                |f(x_{n+1})|
                le L,|x_{n+1}-y_{n+1}|,frac{|f(y_{n+1})|}{|f'(x_n)|}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
                le frac{L^2}{2}frac{|f(x_n)|^3}{|f'(x_n)|^4}+frac{L^3}{8}frac{|f(x_{n+1})|^4}{|f'(x_n)|^6}
                $$

                As the function value stands as proxy for the distance to the root, this shows the third order convergence.





                Note however that in terms of effort, or the Ostrowski index, you only get for polynomial $f$ an order of $sqrt[3]{3}=1.442$ per function and derivative evaluation, in the non-polynomial case even only $sqrt[4]3=1.316$ as one derivative costs as much as 2 function evaluations. To compare, the Newton method has $sqrt2=1.414$ resp. $sqrt[3]2=1.260$ and the secant method $frac{1+sqrt5}2=1.618$ (if it converges).



                So this method is slightly better than the Newton method. To put it in integer terms, in the polynomial case within the effort of $6$ function evaluations one can perform




                • 3 Newton steps with error reduction from $epsilon$ to $ϵ^8$ or

                • 2 cycles of this two-step method with an error reduction to $ϵ^9$.


                In the non-polynomial case, in the time of $12$ evaluations of $f$ one can perform





                • $4$ Newton steps to $ϵ^{16}$ or


                • $3$ two-step cycles to $ϵ^{27}$.







                share|cite|improve this answer














                share|cite|improve this answer



                share|cite|improve this answer








                edited Dec 17 '18 at 16:43

























                answered Dec 16 '18 at 21:21









                LutzLLutzL

                59.6k42057




                59.6k42057






























                    draft saved

                    draft discarded




















































                    Thanks for contributing an answer to Mathematics Stack Exchange!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid



                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.


                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function () {
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3043066%2forder-of-convergence-of-a-two-step-iteration-method%23new-answer', 'question_page');
                    }
                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Plaza Victoria

                    In PowerPoint, is there a keyboard shortcut for bulleted / numbered list?

                    How to put 3 figures in Latex with 2 figures side by side and 1 below these side by side images but in...