Sum of All Combinations of Products of Two Matrices
$begingroup$
Suppose that $mathbf A$ and $mathbf B$ are square, diagonalizable matrices. Consider the following infinite sum of all combinations of these two matrices:
begin{align} mathbf S = mathbf I &+mathbf A + mathbf B +\
&+mathbf A^2 + mathbf Amathbf B + mathbf Bmathbf A + mathbf B^2+ \
&+mathbf A^3 + mathbf A^2mathbf B + mathbf Amathbf Bmathbf A+ mathbf Amathbf B^2 + mathbf Bmathbf A^2+ mathbf Bmathbf Amathbf B+ mathbf B^2mathbf A + mathbf B^3+\
&+cdots
end{align}
Let $mathbf A=mathbf Q_Amathbf Lambda_Amathbf Q_A^{-1}$, and $mathbf B=mathbf Q_Bmathbf Lambda_Bmathbf Q_B^{-1}$ be the eigendecompositions of $mathbf A$ and $mathbf B$.
If $mathbf B=mathbf 0$, then the sum would be equal to
$$mathbf S = sum_{k=0}^infty mathbf A^k=sum_{k=0}^infty(mathbf Q_Amathbf Lambda_Amathbf Q_A^{-1})^k=mathbf Q_A left(sum_{k=0}^inftymathbfLambda_A^kright)mathbf Q_A^{-1}.$$
Is there a general way to express $mathbf S$ as a function of $mathbf Q_A$, $mathbf Q_B$, and sums that involve $mathbf Lambda_A$ and $mathbf Lambda_B$?
linear-algebra combinatorics matrices summation
$endgroup$
add a comment |
$begingroup$
Suppose that $mathbf A$ and $mathbf B$ are square, diagonalizable matrices. Consider the following infinite sum of all combinations of these two matrices:
begin{align} mathbf S = mathbf I &+mathbf A + mathbf B +\
&+mathbf A^2 + mathbf Amathbf B + mathbf Bmathbf A + mathbf B^2+ \
&+mathbf A^3 + mathbf A^2mathbf B + mathbf Amathbf Bmathbf A+ mathbf Amathbf B^2 + mathbf Bmathbf A^2+ mathbf Bmathbf Amathbf B+ mathbf B^2mathbf A + mathbf B^3+\
&+cdots
end{align}
Let $mathbf A=mathbf Q_Amathbf Lambda_Amathbf Q_A^{-1}$, and $mathbf B=mathbf Q_Bmathbf Lambda_Bmathbf Q_B^{-1}$ be the eigendecompositions of $mathbf A$ and $mathbf B$.
If $mathbf B=mathbf 0$, then the sum would be equal to
$$mathbf S = sum_{k=0}^infty mathbf A^k=sum_{k=0}^infty(mathbf Q_Amathbf Lambda_Amathbf Q_A^{-1})^k=mathbf Q_A left(sum_{k=0}^inftymathbfLambda_A^kright)mathbf Q_A^{-1}.$$
Is there a general way to express $mathbf S$ as a function of $mathbf Q_A$, $mathbf Q_B$, and sums that involve $mathbf Lambda_A$ and $mathbf Lambda_B$?
linear-algebra combinatorics matrices summation
$endgroup$
3
$begingroup$
Formally, the sum of all products of length $k$ is $({bf A} + {bf B})^k$. In particular, if ${bf A}, {bf B}$ commute, they are simultaneously diagonalizable, so we can take ${bf Q}_A = {bf Q}_B$, and then $S$ Is given by the same formula as above but replacing ${bf Lambda}_A^k$ with $({bf Lambda}_A + {bf Lambda}_B)^k$.
$endgroup$
– Travis
Dec 12 '18 at 14:24
2
$begingroup$
In general, though, I don't think one can expect a formula in terms of ${bf Q}_A, {bf Q}_B$, and sums that involve ${bf Lambda}_A$ and ${bf Lambda}_B$ but not ${bf Q}_A$ or ${bf Q}_B$, since they contain the data of the relative positions of eigenspaces of ${bf A}, {bf B}$. In general, the eigenvalues of a sum ${bf A} + {bf B}$ of two matrices cannot be determined (completely) from the eigenvalues of ${bf A}$ and ${bf B}$.
$endgroup$
– Travis
Dec 12 '18 at 14:34
$begingroup$
Just a thought: is $S = I + AS + BS$?
$endgroup$
– player100
Dec 12 '18 at 14:40
$begingroup$
following up on @Travis comments: in fact the key is that $S = sum^infty_{k=0} C^k$ where $C = A+B$, and so there is a nice formula depending on the decomposition of $C$. However, there is no guarantee $C$ is diagonalizable, so you might need the Jordan form, etc.
$endgroup$
– antkam
Dec 12 '18 at 16:55
add a comment |
$begingroup$
Suppose that $mathbf A$ and $mathbf B$ are square, diagonalizable matrices. Consider the following infinite sum of all combinations of these two matrices:
begin{align} mathbf S = mathbf I &+mathbf A + mathbf B +\
&+mathbf A^2 + mathbf Amathbf B + mathbf Bmathbf A + mathbf B^2+ \
&+mathbf A^3 + mathbf A^2mathbf B + mathbf Amathbf Bmathbf A+ mathbf Amathbf B^2 + mathbf Bmathbf A^2+ mathbf Bmathbf Amathbf B+ mathbf B^2mathbf A + mathbf B^3+\
&+cdots
end{align}
Let $mathbf A=mathbf Q_Amathbf Lambda_Amathbf Q_A^{-1}$, and $mathbf B=mathbf Q_Bmathbf Lambda_Bmathbf Q_B^{-1}$ be the eigendecompositions of $mathbf A$ and $mathbf B$.
If $mathbf B=mathbf 0$, then the sum would be equal to
$$mathbf S = sum_{k=0}^infty mathbf A^k=sum_{k=0}^infty(mathbf Q_Amathbf Lambda_Amathbf Q_A^{-1})^k=mathbf Q_A left(sum_{k=0}^inftymathbfLambda_A^kright)mathbf Q_A^{-1}.$$
Is there a general way to express $mathbf S$ as a function of $mathbf Q_A$, $mathbf Q_B$, and sums that involve $mathbf Lambda_A$ and $mathbf Lambda_B$?
linear-algebra combinatorics matrices summation
$endgroup$
Suppose that $mathbf A$ and $mathbf B$ are square, diagonalizable matrices. Consider the following infinite sum of all combinations of these two matrices:
begin{align} mathbf S = mathbf I &+mathbf A + mathbf B +\
&+mathbf A^2 + mathbf Amathbf B + mathbf Bmathbf A + mathbf B^2+ \
&+mathbf A^3 + mathbf A^2mathbf B + mathbf Amathbf Bmathbf A+ mathbf Amathbf B^2 + mathbf Bmathbf A^2+ mathbf Bmathbf Amathbf B+ mathbf B^2mathbf A + mathbf B^3+\
&+cdots
end{align}
Let $mathbf A=mathbf Q_Amathbf Lambda_Amathbf Q_A^{-1}$, and $mathbf B=mathbf Q_Bmathbf Lambda_Bmathbf Q_B^{-1}$ be the eigendecompositions of $mathbf A$ and $mathbf B$.
If $mathbf B=mathbf 0$, then the sum would be equal to
$$mathbf S = sum_{k=0}^infty mathbf A^k=sum_{k=0}^infty(mathbf Q_Amathbf Lambda_Amathbf Q_A^{-1})^k=mathbf Q_A left(sum_{k=0}^inftymathbfLambda_A^kright)mathbf Q_A^{-1}.$$
Is there a general way to express $mathbf S$ as a function of $mathbf Q_A$, $mathbf Q_B$, and sums that involve $mathbf Lambda_A$ and $mathbf Lambda_B$?
linear-algebra combinatorics matrices summation
linear-algebra combinatorics matrices summation
edited Dec 12 '18 at 14:09
mzp
asked Dec 12 '18 at 13:56
mzpmzp
1,24721234
1,24721234
3
$begingroup$
Formally, the sum of all products of length $k$ is $({bf A} + {bf B})^k$. In particular, if ${bf A}, {bf B}$ commute, they are simultaneously diagonalizable, so we can take ${bf Q}_A = {bf Q}_B$, and then $S$ Is given by the same formula as above but replacing ${bf Lambda}_A^k$ with $({bf Lambda}_A + {bf Lambda}_B)^k$.
$endgroup$
– Travis
Dec 12 '18 at 14:24
2
$begingroup$
In general, though, I don't think one can expect a formula in terms of ${bf Q}_A, {bf Q}_B$, and sums that involve ${bf Lambda}_A$ and ${bf Lambda}_B$ but not ${bf Q}_A$ or ${bf Q}_B$, since they contain the data of the relative positions of eigenspaces of ${bf A}, {bf B}$. In general, the eigenvalues of a sum ${bf A} + {bf B}$ of two matrices cannot be determined (completely) from the eigenvalues of ${bf A}$ and ${bf B}$.
$endgroup$
– Travis
Dec 12 '18 at 14:34
$begingroup$
Just a thought: is $S = I + AS + BS$?
$endgroup$
– player100
Dec 12 '18 at 14:40
$begingroup$
following up on @Travis comments: in fact the key is that $S = sum^infty_{k=0} C^k$ where $C = A+B$, and so there is a nice formula depending on the decomposition of $C$. However, there is no guarantee $C$ is diagonalizable, so you might need the Jordan form, etc.
$endgroup$
– antkam
Dec 12 '18 at 16:55
add a comment |
3
$begingroup$
Formally, the sum of all products of length $k$ is $({bf A} + {bf B})^k$. In particular, if ${bf A}, {bf B}$ commute, they are simultaneously diagonalizable, so we can take ${bf Q}_A = {bf Q}_B$, and then $S$ Is given by the same formula as above but replacing ${bf Lambda}_A^k$ with $({bf Lambda}_A + {bf Lambda}_B)^k$.
$endgroup$
– Travis
Dec 12 '18 at 14:24
2
$begingroup$
In general, though, I don't think one can expect a formula in terms of ${bf Q}_A, {bf Q}_B$, and sums that involve ${bf Lambda}_A$ and ${bf Lambda}_B$ but not ${bf Q}_A$ or ${bf Q}_B$, since they contain the data of the relative positions of eigenspaces of ${bf A}, {bf B}$. In general, the eigenvalues of a sum ${bf A} + {bf B}$ of two matrices cannot be determined (completely) from the eigenvalues of ${bf A}$ and ${bf B}$.
$endgroup$
– Travis
Dec 12 '18 at 14:34
$begingroup$
Just a thought: is $S = I + AS + BS$?
$endgroup$
– player100
Dec 12 '18 at 14:40
$begingroup$
following up on @Travis comments: in fact the key is that $S = sum^infty_{k=0} C^k$ where $C = A+B$, and so there is a nice formula depending on the decomposition of $C$. However, there is no guarantee $C$ is diagonalizable, so you might need the Jordan form, etc.
$endgroup$
– antkam
Dec 12 '18 at 16:55
3
3
$begingroup$
Formally, the sum of all products of length $k$ is $({bf A} + {bf B})^k$. In particular, if ${bf A}, {bf B}$ commute, they are simultaneously diagonalizable, so we can take ${bf Q}_A = {bf Q}_B$, and then $S$ Is given by the same formula as above but replacing ${bf Lambda}_A^k$ with $({bf Lambda}_A + {bf Lambda}_B)^k$.
$endgroup$
– Travis
Dec 12 '18 at 14:24
$begingroup$
Formally, the sum of all products of length $k$ is $({bf A} + {bf B})^k$. In particular, if ${bf A}, {bf B}$ commute, they are simultaneously diagonalizable, so we can take ${bf Q}_A = {bf Q}_B$, and then $S$ Is given by the same formula as above but replacing ${bf Lambda}_A^k$ with $({bf Lambda}_A + {bf Lambda}_B)^k$.
$endgroup$
– Travis
Dec 12 '18 at 14:24
2
2
$begingroup$
In general, though, I don't think one can expect a formula in terms of ${bf Q}_A, {bf Q}_B$, and sums that involve ${bf Lambda}_A$ and ${bf Lambda}_B$ but not ${bf Q}_A$ or ${bf Q}_B$, since they contain the data of the relative positions of eigenspaces of ${bf A}, {bf B}$. In general, the eigenvalues of a sum ${bf A} + {bf B}$ of two matrices cannot be determined (completely) from the eigenvalues of ${bf A}$ and ${bf B}$.
$endgroup$
– Travis
Dec 12 '18 at 14:34
$begingroup$
In general, though, I don't think one can expect a formula in terms of ${bf Q}_A, {bf Q}_B$, and sums that involve ${bf Lambda}_A$ and ${bf Lambda}_B$ but not ${bf Q}_A$ or ${bf Q}_B$, since they contain the data of the relative positions of eigenspaces of ${bf A}, {bf B}$. In general, the eigenvalues of a sum ${bf A} + {bf B}$ of two matrices cannot be determined (completely) from the eigenvalues of ${bf A}$ and ${bf B}$.
$endgroup$
– Travis
Dec 12 '18 at 14:34
$begingroup$
Just a thought: is $S = I + AS + BS$?
$endgroup$
– player100
Dec 12 '18 at 14:40
$begingroup$
Just a thought: is $S = I + AS + BS$?
$endgroup$
– player100
Dec 12 '18 at 14:40
$begingroup$
following up on @Travis comments: in fact the key is that $S = sum^infty_{k=0} C^k$ where $C = A+B$, and so there is a nice formula depending on the decomposition of $C$. However, there is no guarantee $C$ is diagonalizable, so you might need the Jordan form, etc.
$endgroup$
– antkam
Dec 12 '18 at 16:55
$begingroup$
following up on @Travis comments: in fact the key is that $S = sum^infty_{k=0} C^k$ where $C = A+B$, and so there is a nice formula depending on the decomposition of $C$. However, there is no guarantee $C$ is diagonalizable, so you might need the Jordan form, etc.
$endgroup$
– antkam
Dec 12 '18 at 16:55
add a comment |
0
active
oldest
votes
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3036715%2fsum-of-all-combinations-of-products-of-two-matrices%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
0
active
oldest
votes
0
active
oldest
votes
active
oldest
votes
active
oldest
votes
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3036715%2fsum-of-all-combinations-of-products-of-two-matrices%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
3
$begingroup$
Formally, the sum of all products of length $k$ is $({bf A} + {bf B})^k$. In particular, if ${bf A}, {bf B}$ commute, they are simultaneously diagonalizable, so we can take ${bf Q}_A = {bf Q}_B$, and then $S$ Is given by the same formula as above but replacing ${bf Lambda}_A^k$ with $({bf Lambda}_A + {bf Lambda}_B)^k$.
$endgroup$
– Travis
Dec 12 '18 at 14:24
2
$begingroup$
In general, though, I don't think one can expect a formula in terms of ${bf Q}_A, {bf Q}_B$, and sums that involve ${bf Lambda}_A$ and ${bf Lambda}_B$ but not ${bf Q}_A$ or ${bf Q}_B$, since they contain the data of the relative positions of eigenspaces of ${bf A}, {bf B}$. In general, the eigenvalues of a sum ${bf A} + {bf B}$ of two matrices cannot be determined (completely) from the eigenvalues of ${bf A}$ and ${bf B}$.
$endgroup$
– Travis
Dec 12 '18 at 14:34
$begingroup$
Just a thought: is $S = I + AS + BS$?
$endgroup$
– player100
Dec 12 '18 at 14:40
$begingroup$
following up on @Travis comments: in fact the key is that $S = sum^infty_{k=0} C^k$ where $C = A+B$, and so there is a nice formula depending on the decomposition of $C$. However, there is no guarantee $C$ is diagonalizable, so you might need the Jordan form, etc.
$endgroup$
– antkam
Dec 12 '18 at 16:55