What are the complex roots of $x^3-1$?











up vote
2
down vote

favorite












What are the complex roots of $x^3-1$?



Work I've done so far:



I've set $x = a + bi$. Since $x^3-1=0$, I set $x^3 = (a+bi)^3=1$.



This gives me the following:



(1) $(-ab^2 + a^3) + (2ab^2 + 2a^2b + a^2b - b^3)i$



Which means that I set $(-ab^2 + a^3) = a(a^2-b^2)= 1$ which is also equivalent to



(2) $a(a-b)(a+b)=(a-b)(a^2+ab)=1$.



I also set



(3) $(2ab^2 + 2a^2b + a^2b - b^3) = 0$.



I simplify (3) to



(4) $2b(a^2 + ab) + (a^2 -b^2)b = 0 $



which gives me



(5) $frac{2b}{a-b} + frac{b}{a} = 0$ using (2).



Then I get



(6) $frac{2b}{a-b} = -frac{b}{a}$. Then I get that $3a=b$. Plugging into (2) I get



(7) $a(a^2 - (3a)^2)=1 = a(a^2 -9a^2) = -8a^3$. So that $a= frac{-1}{2}$. Now I get that $b= frac{3}{2}$, which would give me $frac{1}{2} + frac{3}{2}i$. But on Wolfram, the imaginary component is close to $.9$. Where am I going wrong?










share|cite|improve this question




















  • 4




    (1) is already erroneous.
    – Lord Shark the Unknown
    Dec 2 at 20:46






  • 1




    It would be better to factor $x^3-1$ first.
    – Abraham Zhang
    Dec 2 at 20:47










  • @LordSharktheUnknown: Is the algebra incorrect?
    – K.M
    Dec 2 at 20:52






  • 1




    @K.M Yes, your algebraic manipulations are incorrect. Use Pascal's triangle to see where your binomial expansion went wrong.
    – Scounged
    Dec 2 at 20:54















up vote
2
down vote

favorite












What are the complex roots of $x^3-1$?



Work I've done so far:



I've set $x = a + bi$. Since $x^3-1=0$, I set $x^3 = (a+bi)^3=1$.



This gives me the following:



(1) $(-ab^2 + a^3) + (2ab^2 + 2a^2b + a^2b - b^3)i$



Which means that I set $(-ab^2 + a^3) = a(a^2-b^2)= 1$ which is also equivalent to



(2) $a(a-b)(a+b)=(a-b)(a^2+ab)=1$.



I also set



(3) $(2ab^2 + 2a^2b + a^2b - b^3) = 0$.



I simplify (3) to



(4) $2b(a^2 + ab) + (a^2 -b^2)b = 0 $



which gives me



(5) $frac{2b}{a-b} + frac{b}{a} = 0$ using (2).



Then I get



(6) $frac{2b}{a-b} = -frac{b}{a}$. Then I get that $3a=b$. Plugging into (2) I get



(7) $a(a^2 - (3a)^2)=1 = a(a^2 -9a^2) = -8a^3$. So that $a= frac{-1}{2}$. Now I get that $b= frac{3}{2}$, which would give me $frac{1}{2} + frac{3}{2}i$. But on Wolfram, the imaginary component is close to $.9$. Where am I going wrong?










share|cite|improve this question




















  • 4




    (1) is already erroneous.
    – Lord Shark the Unknown
    Dec 2 at 20:46






  • 1




    It would be better to factor $x^3-1$ first.
    – Abraham Zhang
    Dec 2 at 20:47










  • @LordSharktheUnknown: Is the algebra incorrect?
    – K.M
    Dec 2 at 20:52






  • 1




    @K.M Yes, your algebraic manipulations are incorrect. Use Pascal's triangle to see where your binomial expansion went wrong.
    – Scounged
    Dec 2 at 20:54













up vote
2
down vote

favorite









up vote
2
down vote

favorite











What are the complex roots of $x^3-1$?



Work I've done so far:



I've set $x = a + bi$. Since $x^3-1=0$, I set $x^3 = (a+bi)^3=1$.



This gives me the following:



(1) $(-ab^2 + a^3) + (2ab^2 + 2a^2b + a^2b - b^3)i$



Which means that I set $(-ab^2 + a^3) = a(a^2-b^2)= 1$ which is also equivalent to



(2) $a(a-b)(a+b)=(a-b)(a^2+ab)=1$.



I also set



(3) $(2ab^2 + 2a^2b + a^2b - b^3) = 0$.



I simplify (3) to



(4) $2b(a^2 + ab) + (a^2 -b^2)b = 0 $



which gives me



(5) $frac{2b}{a-b} + frac{b}{a} = 0$ using (2).



Then I get



(6) $frac{2b}{a-b} = -frac{b}{a}$. Then I get that $3a=b$. Plugging into (2) I get



(7) $a(a^2 - (3a)^2)=1 = a(a^2 -9a^2) = -8a^3$. So that $a= frac{-1}{2}$. Now I get that $b= frac{3}{2}$, which would give me $frac{1}{2} + frac{3}{2}i$. But on Wolfram, the imaginary component is close to $.9$. Where am I going wrong?










share|cite|improve this question















What are the complex roots of $x^3-1$?



Work I've done so far:



I've set $x = a + bi$. Since $x^3-1=0$, I set $x^3 = (a+bi)^3=1$.



This gives me the following:



(1) $(-ab^2 + a^3) + (2ab^2 + 2a^2b + a^2b - b^3)i$



Which means that I set $(-ab^2 + a^3) = a(a^2-b^2)= 1$ which is also equivalent to



(2) $a(a-b)(a+b)=(a-b)(a^2+ab)=1$.



I also set



(3) $(2ab^2 + 2a^2b + a^2b - b^3) = 0$.



I simplify (3) to



(4) $2b(a^2 + ab) + (a^2 -b^2)b = 0 $



which gives me



(5) $frac{2b}{a-b} + frac{b}{a} = 0$ using (2).



Then I get



(6) $frac{2b}{a-b} = -frac{b}{a}$. Then I get that $3a=b$. Plugging into (2) I get



(7) $a(a^2 - (3a)^2)=1 = a(a^2 -9a^2) = -8a^3$. So that $a= frac{-1}{2}$. Now I get that $b= frac{3}{2}$, which would give me $frac{1}{2} + frac{3}{2}i$. But on Wolfram, the imaginary component is close to $.9$. Where am I going wrong?







complex-analysis






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Dec 2 at 20:56

























asked Dec 2 at 20:44









K.M

651312




651312








  • 4




    (1) is already erroneous.
    – Lord Shark the Unknown
    Dec 2 at 20:46






  • 1




    It would be better to factor $x^3-1$ first.
    – Abraham Zhang
    Dec 2 at 20:47










  • @LordSharktheUnknown: Is the algebra incorrect?
    – K.M
    Dec 2 at 20:52






  • 1




    @K.M Yes, your algebraic manipulations are incorrect. Use Pascal's triangle to see where your binomial expansion went wrong.
    – Scounged
    Dec 2 at 20:54














  • 4




    (1) is already erroneous.
    – Lord Shark the Unknown
    Dec 2 at 20:46






  • 1




    It would be better to factor $x^3-1$ first.
    – Abraham Zhang
    Dec 2 at 20:47










  • @LordSharktheUnknown: Is the algebra incorrect?
    – K.M
    Dec 2 at 20:52






  • 1




    @K.M Yes, your algebraic manipulations are incorrect. Use Pascal's triangle to see where your binomial expansion went wrong.
    – Scounged
    Dec 2 at 20:54








4




4




(1) is already erroneous.
– Lord Shark the Unknown
Dec 2 at 20:46




(1) is already erroneous.
– Lord Shark the Unknown
Dec 2 at 20:46




1




1




It would be better to factor $x^3-1$ first.
– Abraham Zhang
Dec 2 at 20:47




It would be better to factor $x^3-1$ first.
– Abraham Zhang
Dec 2 at 20:47












@LordSharktheUnknown: Is the algebra incorrect?
– K.M
Dec 2 at 20:52




@LordSharktheUnknown: Is the algebra incorrect?
– K.M
Dec 2 at 20:52




1




1




@K.M Yes, your algebraic manipulations are incorrect. Use Pascal's triangle to see where your binomial expansion went wrong.
– Scounged
Dec 2 at 20:54




@K.M Yes, your algebraic manipulations are incorrect. Use Pascal's triangle to see where your binomial expansion went wrong.
– Scounged
Dec 2 at 20:54










2 Answers
2






active

oldest

votes

















up vote
6
down vote



accepted










Hint:



The simpler way is to factorize:
$$
x^3-1=(x-1)(x^2+x+1)
$$



can you find all the roots?





Anyway, your algebra is wrong because:



$$
(a+ib)^3=a^3+3a^2(ib)+3a(ib)^2+(ib)^3=a^3-3ab^2+i(3a^2b-b^3)
$$






share|cite|improve this answer



















  • 1




    Thank you for not just giving the hint about an easier solution technique, but also for identifying exactly where the asker has gone astray.
    – T. Bongers
    Dec 2 at 21:06


















up vote
2
down vote













You can just use difference of cubes, which gets the answer much more quickly:



$$a^3-b^3 = (a-b)(a^2+ab+b^2)$$



Applying it here, you get



$$x^3-1 = (x-1)(x^2+x+1) = 0$$



The first factor obviously gives the real root of $x = 1$, so solve for the second factor. It should be pretty straightforward.





As for your error, you have expanded incorrectly in $(1)$. Recall that



$$(a+b)^3 = a^3+3a^2b+3ab^2+b^3$$



and you apply it to



$$(a+bi)^3$$






share|cite|improve this answer























    Your Answer





    StackExchange.ifUsing("editor", function () {
    return StackExchange.using("mathjaxEditing", function () {
    StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
    StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
    });
    });
    }, "mathjax-editing");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "69"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    noCode: true, onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3023153%2fwhat-are-the-complex-roots-of-x3-1%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    2 Answers
    2






    active

    oldest

    votes








    2 Answers
    2






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes








    up vote
    6
    down vote



    accepted










    Hint:



    The simpler way is to factorize:
    $$
    x^3-1=(x-1)(x^2+x+1)
    $$



    can you find all the roots?





    Anyway, your algebra is wrong because:



    $$
    (a+ib)^3=a^3+3a^2(ib)+3a(ib)^2+(ib)^3=a^3-3ab^2+i(3a^2b-b^3)
    $$






    share|cite|improve this answer



















    • 1




      Thank you for not just giving the hint about an easier solution technique, but also for identifying exactly where the asker has gone astray.
      – T. Bongers
      Dec 2 at 21:06















    up vote
    6
    down vote



    accepted










    Hint:



    The simpler way is to factorize:
    $$
    x^3-1=(x-1)(x^2+x+1)
    $$



    can you find all the roots?





    Anyway, your algebra is wrong because:



    $$
    (a+ib)^3=a^3+3a^2(ib)+3a(ib)^2+(ib)^3=a^3-3ab^2+i(3a^2b-b^3)
    $$






    share|cite|improve this answer



















    • 1




      Thank you for not just giving the hint about an easier solution technique, but also for identifying exactly where the asker has gone astray.
      – T. Bongers
      Dec 2 at 21:06













    up vote
    6
    down vote



    accepted







    up vote
    6
    down vote



    accepted






    Hint:



    The simpler way is to factorize:
    $$
    x^3-1=(x-1)(x^2+x+1)
    $$



    can you find all the roots?





    Anyway, your algebra is wrong because:



    $$
    (a+ib)^3=a^3+3a^2(ib)+3a(ib)^2+(ib)^3=a^3-3ab^2+i(3a^2b-b^3)
    $$






    share|cite|improve this answer














    Hint:



    The simpler way is to factorize:
    $$
    x^3-1=(x-1)(x^2+x+1)
    $$



    can you find all the roots?





    Anyway, your algebra is wrong because:



    $$
    (a+ib)^3=a^3+3a^2(ib)+3a(ib)^2+(ib)^3=a^3-3ab^2+i(3a^2b-b^3)
    $$







    share|cite|improve this answer














    share|cite|improve this answer



    share|cite|improve this answer








    edited Dec 2 at 21:03

























    answered Dec 2 at 20:47









    Emilio Novati

    51.2k43472




    51.2k43472








    • 1




      Thank you for not just giving the hint about an easier solution technique, but also for identifying exactly where the asker has gone astray.
      – T. Bongers
      Dec 2 at 21:06














    • 1




      Thank you for not just giving the hint about an easier solution technique, but also for identifying exactly where the asker has gone astray.
      – T. Bongers
      Dec 2 at 21:06








    1




    1




    Thank you for not just giving the hint about an easier solution technique, but also for identifying exactly where the asker has gone astray.
    – T. Bongers
    Dec 2 at 21:06




    Thank you for not just giving the hint about an easier solution technique, but also for identifying exactly where the asker has gone astray.
    – T. Bongers
    Dec 2 at 21:06










    up vote
    2
    down vote













    You can just use difference of cubes, which gets the answer much more quickly:



    $$a^3-b^3 = (a-b)(a^2+ab+b^2)$$



    Applying it here, you get



    $$x^3-1 = (x-1)(x^2+x+1) = 0$$



    The first factor obviously gives the real root of $x = 1$, so solve for the second factor. It should be pretty straightforward.





    As for your error, you have expanded incorrectly in $(1)$. Recall that



    $$(a+b)^3 = a^3+3a^2b+3ab^2+b^3$$



    and you apply it to



    $$(a+bi)^3$$






    share|cite|improve this answer



























      up vote
      2
      down vote













      You can just use difference of cubes, which gets the answer much more quickly:



      $$a^3-b^3 = (a-b)(a^2+ab+b^2)$$



      Applying it here, you get



      $$x^3-1 = (x-1)(x^2+x+1) = 0$$



      The first factor obviously gives the real root of $x = 1$, so solve for the second factor. It should be pretty straightforward.





      As for your error, you have expanded incorrectly in $(1)$. Recall that



      $$(a+b)^3 = a^3+3a^2b+3ab^2+b^3$$



      and you apply it to



      $$(a+bi)^3$$






      share|cite|improve this answer

























        up vote
        2
        down vote










        up vote
        2
        down vote









        You can just use difference of cubes, which gets the answer much more quickly:



        $$a^3-b^3 = (a-b)(a^2+ab+b^2)$$



        Applying it here, you get



        $$x^3-1 = (x-1)(x^2+x+1) = 0$$



        The first factor obviously gives the real root of $x = 1$, so solve for the second factor. It should be pretty straightforward.





        As for your error, you have expanded incorrectly in $(1)$. Recall that



        $$(a+b)^3 = a^3+3a^2b+3ab^2+b^3$$



        and you apply it to



        $$(a+bi)^3$$






        share|cite|improve this answer














        You can just use difference of cubes, which gets the answer much more quickly:



        $$a^3-b^3 = (a-b)(a^2+ab+b^2)$$



        Applying it here, you get



        $$x^3-1 = (x-1)(x^2+x+1) = 0$$



        The first factor obviously gives the real root of $x = 1$, so solve for the second factor. It should be pretty straightforward.





        As for your error, you have expanded incorrectly in $(1)$. Recall that



        $$(a+b)^3 = a^3+3a^2b+3ab^2+b^3$$



        and you apply it to



        $$(a+bi)^3$$







        share|cite|improve this answer














        share|cite|improve this answer



        share|cite|improve this answer








        edited Dec 2 at 21:12

























        answered Dec 2 at 21:00









        KM101

        3,466417




        3,466417






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Mathematics Stack Exchange!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.





            Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


            Please pay close attention to the following guidance:


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3023153%2fwhat-are-the-complex-roots-of-x3-1%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            Bressuire

            Cabo Verde

            Gyllenstierna