Kernel of matrix polynomials The Next CEO of Stack OverflowLinear Algebra: Minimum Polynomial...

Do I need to write [sic] when a number is less than 10 but isn't written out?

How to delete every two lines after 3rd lines in a file contains very large number of lines?

Is French Guiana a (hard) EU border?

How to edit “Name” property in GCI output?

TikZ: How to reverse arrow direction without switching start/end point?

The past simple of "gaslight" – "gaslighted" or "gaslit"?

Why is the US ranked as #45 in Press Freedom ratings, despite its extremely permissive free speech laws?

Help understanding this unsettling image of Titan, Epimetheus, and Saturn's rings?

Would a completely good Muggle be able to use a wand?

Bartok - Syncopation (1): Meaning of notes in between Grand Staff

How to write a definition with variants?

Can this equation be simplified further?

Why the difference in type-inference over the as-pattern in two similar function definitions?

Rotate a column

Is it possible to replace duplicates of a character with one character using tr

Is it my responsibility to learn a new technology in my own time my employer wants to implement?

Is micro rebar a better way to reinforce concrete than rebar?

Won the lottery - how do I keep the money?

Method for adding error messages to a dictionary given a key

Is it okay to majorly distort historical facts while writing a fiction story?

What did we know about the Kessel run before the prequels?

When you upcast Blindness/Deafness, do all targets suffer the same effect?

Writing differences on a blackboard

Running a General Election and the European Elections together



Kernel of matrix polynomials



The Next CEO of Stack OverflowLinear Algebra: Minimum Polynomial QuestionShowing diagonalisability using primary decompositionshow T is one one iff $ker(T)=0$Kernel of GCD between minimal polynomial and any polynomialSymmetric positive definite matrix, why $text{Im}(S^{-1}A) = (text{Ker}(A))^{bot_S}$?The kernel of nilpotent MatrixIf $P_1$ and $P_2$ are coprime, how do I show that $ker ( (P_1 times P_2 )(f) )=ker(P_1 (f)) oplus ker(P_2(f))$?Understanding sets of polynomials as subspacesChange of basis of the kernel of a rectangular matrixKernel of polynomial of matrix












0












$begingroup$


I am told that $f(A)=g(A)h(A), f(A)=0$ is a polynomial of an $n times n$ matrix $A$.



I am further told that $g(x)$ and $h(x)$ are coprime. I need to show that $mathbb{R}^n$ is the direct sum of $text{ker }g(A)$ and $text{ker }h(A)$.



My attempt at a solution:
Since $g(x)$ and $h(x)$ are coprime, this means that they have no roots or factors in common, so I can deduce that $text{ker }g(A) cap text{ker }h(A) ={0}$.



I now want to show that if a vector $v in mathbb{R}^n$ is not in $text{ker }g(A)$, then it is in $text{ker }h(A)$ and vice versa.



Since $f(A)=0$, we have $f(A)v=0 Rightarrow g(A)h(A)v=0$



Because of commutativity of polynomial multiplication, this also means that $h(A)g(A)v=0$.



If $v notin text{ker }g(A)$ then $g(A)vne 0$ so $h(A)v=0$. Is this true? Am I on the right track?



Any help would be greatly appreciated!










share|cite|improve this question









$endgroup$








  • 1




    $begingroup$
    You target is wrong : the direct sum is not the disjoint union. You need to show that any $v$ can be written as $v_1 + v_2$ where $v_1 in ker g(A)$ and $v_2 in ker h(A)$. For this purpose you must use relatively primality.
    $endgroup$
    – астон вілла олоф мэллбэрг
    Mar 17 at 7:46












  • $begingroup$
    Thanks. Do you have any hints on how to do this? All I can think of is showing that since $f(A)v=0 forall v in mathbb{R}^n$ then $f(A)(v_1+v_2)=0$ so $v_1+v_2$ must be in one of the kernels - but can't be in both.
    $endgroup$
    – Andrew
    Mar 17 at 7:50










  • $begingroup$
    @Andrew Hint: Bezout's identity.
    $endgroup$
    – M. Vinay
    Mar 17 at 7:52


















0












$begingroup$


I am told that $f(A)=g(A)h(A), f(A)=0$ is a polynomial of an $n times n$ matrix $A$.



I am further told that $g(x)$ and $h(x)$ are coprime. I need to show that $mathbb{R}^n$ is the direct sum of $text{ker }g(A)$ and $text{ker }h(A)$.



My attempt at a solution:
Since $g(x)$ and $h(x)$ are coprime, this means that they have no roots or factors in common, so I can deduce that $text{ker }g(A) cap text{ker }h(A) ={0}$.



I now want to show that if a vector $v in mathbb{R}^n$ is not in $text{ker }g(A)$, then it is in $text{ker }h(A)$ and vice versa.



Since $f(A)=0$, we have $f(A)v=0 Rightarrow g(A)h(A)v=0$



Because of commutativity of polynomial multiplication, this also means that $h(A)g(A)v=0$.



If $v notin text{ker }g(A)$ then $g(A)vne 0$ so $h(A)v=0$. Is this true? Am I on the right track?



Any help would be greatly appreciated!










share|cite|improve this question









$endgroup$








  • 1




    $begingroup$
    You target is wrong : the direct sum is not the disjoint union. You need to show that any $v$ can be written as $v_1 + v_2$ where $v_1 in ker g(A)$ and $v_2 in ker h(A)$. For this purpose you must use relatively primality.
    $endgroup$
    – астон вілла олоф мэллбэрг
    Mar 17 at 7:46












  • $begingroup$
    Thanks. Do you have any hints on how to do this? All I can think of is showing that since $f(A)v=0 forall v in mathbb{R}^n$ then $f(A)(v_1+v_2)=0$ so $v_1+v_2$ must be in one of the kernels - but can't be in both.
    $endgroup$
    – Andrew
    Mar 17 at 7:50










  • $begingroup$
    @Andrew Hint: Bezout's identity.
    $endgroup$
    – M. Vinay
    Mar 17 at 7:52
















0












0








0





$begingroup$


I am told that $f(A)=g(A)h(A), f(A)=0$ is a polynomial of an $n times n$ matrix $A$.



I am further told that $g(x)$ and $h(x)$ are coprime. I need to show that $mathbb{R}^n$ is the direct sum of $text{ker }g(A)$ and $text{ker }h(A)$.



My attempt at a solution:
Since $g(x)$ and $h(x)$ are coprime, this means that they have no roots or factors in common, so I can deduce that $text{ker }g(A) cap text{ker }h(A) ={0}$.



I now want to show that if a vector $v in mathbb{R}^n$ is not in $text{ker }g(A)$, then it is in $text{ker }h(A)$ and vice versa.



Since $f(A)=0$, we have $f(A)v=0 Rightarrow g(A)h(A)v=0$



Because of commutativity of polynomial multiplication, this also means that $h(A)g(A)v=0$.



If $v notin text{ker }g(A)$ then $g(A)vne 0$ so $h(A)v=0$. Is this true? Am I on the right track?



Any help would be greatly appreciated!










share|cite|improve this question









$endgroup$




I am told that $f(A)=g(A)h(A), f(A)=0$ is a polynomial of an $n times n$ matrix $A$.



I am further told that $g(x)$ and $h(x)$ are coprime. I need to show that $mathbb{R}^n$ is the direct sum of $text{ker }g(A)$ and $text{ker }h(A)$.



My attempt at a solution:
Since $g(x)$ and $h(x)$ are coprime, this means that they have no roots or factors in common, so I can deduce that $text{ker }g(A) cap text{ker }h(A) ={0}$.



I now want to show that if a vector $v in mathbb{R}^n$ is not in $text{ker }g(A)$, then it is in $text{ker }h(A)$ and vice versa.



Since $f(A)=0$, we have $f(A)v=0 Rightarrow g(A)h(A)v=0$



Because of commutativity of polynomial multiplication, this also means that $h(A)g(A)v=0$.



If $v notin text{ker }g(A)$ then $g(A)vne 0$ so $h(A)v=0$. Is this true? Am I on the right track?



Any help would be greatly appreciated!







linear-algebra matrices polynomials






share|cite|improve this question













share|cite|improve this question











share|cite|improve this question




share|cite|improve this question










asked Mar 17 at 7:33









AndrewAndrew

357213




357213








  • 1




    $begingroup$
    You target is wrong : the direct sum is not the disjoint union. You need to show that any $v$ can be written as $v_1 + v_2$ where $v_1 in ker g(A)$ and $v_2 in ker h(A)$. For this purpose you must use relatively primality.
    $endgroup$
    – астон вілла олоф мэллбэрг
    Mar 17 at 7:46












  • $begingroup$
    Thanks. Do you have any hints on how to do this? All I can think of is showing that since $f(A)v=0 forall v in mathbb{R}^n$ then $f(A)(v_1+v_2)=0$ so $v_1+v_2$ must be in one of the kernels - but can't be in both.
    $endgroup$
    – Andrew
    Mar 17 at 7:50










  • $begingroup$
    @Andrew Hint: Bezout's identity.
    $endgroup$
    – M. Vinay
    Mar 17 at 7:52
















  • 1




    $begingroup$
    You target is wrong : the direct sum is not the disjoint union. You need to show that any $v$ can be written as $v_1 + v_2$ where $v_1 in ker g(A)$ and $v_2 in ker h(A)$. For this purpose you must use relatively primality.
    $endgroup$
    – астон вілла олоф мэллбэрг
    Mar 17 at 7:46












  • $begingroup$
    Thanks. Do you have any hints on how to do this? All I can think of is showing that since $f(A)v=0 forall v in mathbb{R}^n$ then $f(A)(v_1+v_2)=0$ so $v_1+v_2$ must be in one of the kernels - but can't be in both.
    $endgroup$
    – Andrew
    Mar 17 at 7:50










  • $begingroup$
    @Andrew Hint: Bezout's identity.
    $endgroup$
    – M. Vinay
    Mar 17 at 7:52










1




1




$begingroup$
You target is wrong : the direct sum is not the disjoint union. You need to show that any $v$ can be written as $v_1 + v_2$ where $v_1 in ker g(A)$ and $v_2 in ker h(A)$. For this purpose you must use relatively primality.
$endgroup$
– астон вілла олоф мэллбэрг
Mar 17 at 7:46






$begingroup$
You target is wrong : the direct sum is not the disjoint union. You need to show that any $v$ can be written as $v_1 + v_2$ where $v_1 in ker g(A)$ and $v_2 in ker h(A)$. For this purpose you must use relatively primality.
$endgroup$
– астон вілла олоф мэллбэрг
Mar 17 at 7:46














$begingroup$
Thanks. Do you have any hints on how to do this? All I can think of is showing that since $f(A)v=0 forall v in mathbb{R}^n$ then $f(A)(v_1+v_2)=0$ so $v_1+v_2$ must be in one of the kernels - but can't be in both.
$endgroup$
– Andrew
Mar 17 at 7:50




$begingroup$
Thanks. Do you have any hints on how to do this? All I can think of is showing that since $f(A)v=0 forall v in mathbb{R}^n$ then $f(A)(v_1+v_2)=0$ so $v_1+v_2$ must be in one of the kernels - but can't be in both.
$endgroup$
– Andrew
Mar 17 at 7:50












$begingroup$
@Andrew Hint: Bezout's identity.
$endgroup$
– M. Vinay
Mar 17 at 7:52






$begingroup$
@Andrew Hint: Bezout's identity.
$endgroup$
– M. Vinay
Mar 17 at 7:52












1 Answer
1






active

oldest

votes


















3












$begingroup$

Use the fact that if $g,h$ are relatively prime polynomials then there exist polynomials $p(x),q(x)$ such that $p(x)g(x) + q(x)h(x) = 1$ for all $x$,(Bezout identity), so in particular, $p(A)g(A) + q(A)h(A) = I$.



Now, with this in mind, let $v in Bbb R^n$. Then, we may write $v = p(A)g(A)v + q(A)h(A)v$.



Now, it is clear(by commutativity of the polynomial operators in $A$) that :
$$
h(A)[p(A)g(A)v] = p(A)[g(A)h(A)v] = 0 \
g(A)[q(A)h(A)v] = q(A)[g(A)h(A)v] = 0
$$



Therefore $v$ is a sum of two elements lying in the kernels of $h(A)$ and $g(A)$ respectively. Adding the disjointness of these subspaces gives the required conclusion.






share|cite|improve this answer









$endgroup$














    Your Answer





    StackExchange.ifUsing("editor", function () {
    return StackExchange.using("mathjaxEditing", function () {
    StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
    StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
    });
    });
    }, "mathjax-editing");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "69"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    noCode: true, onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3151238%2fkernel-of-matrix-polynomials%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    3












    $begingroup$

    Use the fact that if $g,h$ are relatively prime polynomials then there exist polynomials $p(x),q(x)$ such that $p(x)g(x) + q(x)h(x) = 1$ for all $x$,(Bezout identity), so in particular, $p(A)g(A) + q(A)h(A) = I$.



    Now, with this in mind, let $v in Bbb R^n$. Then, we may write $v = p(A)g(A)v + q(A)h(A)v$.



    Now, it is clear(by commutativity of the polynomial operators in $A$) that :
    $$
    h(A)[p(A)g(A)v] = p(A)[g(A)h(A)v] = 0 \
    g(A)[q(A)h(A)v] = q(A)[g(A)h(A)v] = 0
    $$



    Therefore $v$ is a sum of two elements lying in the kernels of $h(A)$ and $g(A)$ respectively. Adding the disjointness of these subspaces gives the required conclusion.






    share|cite|improve this answer









    $endgroup$


















      3












      $begingroup$

      Use the fact that if $g,h$ are relatively prime polynomials then there exist polynomials $p(x),q(x)$ such that $p(x)g(x) + q(x)h(x) = 1$ for all $x$,(Bezout identity), so in particular, $p(A)g(A) + q(A)h(A) = I$.



      Now, with this in mind, let $v in Bbb R^n$. Then, we may write $v = p(A)g(A)v + q(A)h(A)v$.



      Now, it is clear(by commutativity of the polynomial operators in $A$) that :
      $$
      h(A)[p(A)g(A)v] = p(A)[g(A)h(A)v] = 0 \
      g(A)[q(A)h(A)v] = q(A)[g(A)h(A)v] = 0
      $$



      Therefore $v$ is a sum of two elements lying in the kernels of $h(A)$ and $g(A)$ respectively. Adding the disjointness of these subspaces gives the required conclusion.






      share|cite|improve this answer









      $endgroup$
















        3












        3








        3





        $begingroup$

        Use the fact that if $g,h$ are relatively prime polynomials then there exist polynomials $p(x),q(x)$ such that $p(x)g(x) + q(x)h(x) = 1$ for all $x$,(Bezout identity), so in particular, $p(A)g(A) + q(A)h(A) = I$.



        Now, with this in mind, let $v in Bbb R^n$. Then, we may write $v = p(A)g(A)v + q(A)h(A)v$.



        Now, it is clear(by commutativity of the polynomial operators in $A$) that :
        $$
        h(A)[p(A)g(A)v] = p(A)[g(A)h(A)v] = 0 \
        g(A)[q(A)h(A)v] = q(A)[g(A)h(A)v] = 0
        $$



        Therefore $v$ is a sum of two elements lying in the kernels of $h(A)$ and $g(A)$ respectively. Adding the disjointness of these subspaces gives the required conclusion.






        share|cite|improve this answer









        $endgroup$



        Use the fact that if $g,h$ are relatively prime polynomials then there exist polynomials $p(x),q(x)$ such that $p(x)g(x) + q(x)h(x) = 1$ for all $x$,(Bezout identity), so in particular, $p(A)g(A) + q(A)h(A) = I$.



        Now, with this in mind, let $v in Bbb R^n$. Then, we may write $v = p(A)g(A)v + q(A)h(A)v$.



        Now, it is clear(by commutativity of the polynomial operators in $A$) that :
        $$
        h(A)[p(A)g(A)v] = p(A)[g(A)h(A)v] = 0 \
        g(A)[q(A)h(A)v] = q(A)[g(A)h(A)v] = 0
        $$



        Therefore $v$ is a sum of two elements lying in the kernels of $h(A)$ and $g(A)$ respectively. Adding the disjointness of these subspaces gives the required conclusion.







        share|cite|improve this answer












        share|cite|improve this answer



        share|cite|improve this answer










        answered Mar 17 at 8:12









        астон вілла олоф мэллбэргастон вілла олоф мэллбэрг

        40.1k33577




        40.1k33577






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Mathematics Stack Exchange!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3151238%2fkernel-of-matrix-polynomials%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            Magento 2 - Add success message with knockout Planned maintenance scheduled April 23, 2019 at 23:30 UTC (7:30pm US/Eastern) Announcing the arrival of Valued Associate #679: Cesar Manara Unicorn Meta Zoo #1: Why another podcast?Success / Error message on ajax request$.widget is not a function when loading a homepage after add custom jQuery on custom themeHow can bind jQuery to current document in Magento 2 When template load by ajaxRedirect page using plugin in Magento 2Magento 2 - Update quantity and totals of cart page without page reload?Magento 2: Quote data not loaded on knockout checkoutMagento 2 : I need to change add to cart success message after adding product into cart through pluginMagento 2.2.5 How to add additional products to cart from new checkout step?Magento 2 Add error/success message with knockoutCan't validate Post Code on checkout page

            Fil:Tokke komm.svg

            Where did Arya get these scars? Unicorn Meta Zoo #1: Why another podcast? Announcing the arrival of Valued Associate #679: Cesar Manara Favourite questions and answers from the 1st quarter of 2019Why did Arya refuse to end it?Has the pronunciation of Arya Stark's name changed?Has Arya forgiven people?Why did Arya Stark lose her vision?Why can Arya still use the faces?Has the Narrow Sea become narrower?Does Arya Stark know how to make poisons outside of the House of Black and White?Why did Nymeria leave Arya?Why did Arya not kill the Lannister soldiers she encountered in the Riverlands?What is the current canonical age of Sansa, Bran and Arya Stark?