Geometric or matrix intuition on $A(A + B)^-1B = B (A + B)^-1 A$
Clash Royale CLAN TAG#URR8PPP
up vote
5
down vote
favorite
I am curious about a seemingly simple identity in matrix algebra. Though matrix multiplication is not commutative (the classic example of noncommutativity, it does allow a commutativity of sorts around a very specific third matrix:
$$
colorblueA (A + B)^-1 B = colorblueB (A + B)^-1 A
$$
There is a very simple algebraic proof:
$$
begineqnarray
colorblueA (A + B)^-1 B + colorredB (A + B)^-1 B &=& (A + B)(A + B)^-1 B\
&=& B\
&=& B (A + B)^-1 (A + B) \
&=& colorblueB (A + B)^-1 A + colorredB (A + B)^-1 B\
endeqnarray
$$
because matrix addition is commutative. The identity follows by cancelling $B (A + B)^-1 B$. (If there is a simpler linear proof, say, without needing cancellation, please say so. That extra weird matrix, while obviously looking right and obviously doing the job, just pops out of nowhere. Or does it?)
Algebra is blind manipulation of symbols. The identity holds in any abstract ring with multiplicative inverses. But for a given model, the identity says something about that model. I've only seen the identity in the context of matrices, but I don't see what's so special about it there.
What does this identity do for matrices?
Is there something, in matrix theory, for which this is special? Does it ever really come up in proofs? $(A+B)^-1$ can't be the only matrices that allow such quasi-commutativity, can they? Is there a visualization or geometric interpretation or a meaningful anything interpretation?
intuition matrix-equations geometric-interpretation
 |Â
show 5 more comments
up vote
5
down vote
favorite
I am curious about a seemingly simple identity in matrix algebra. Though matrix multiplication is not commutative (the classic example of noncommutativity, it does allow a commutativity of sorts around a very specific third matrix:
$$
colorblueA (A + B)^-1 B = colorblueB (A + B)^-1 A
$$
There is a very simple algebraic proof:
$$
begineqnarray
colorblueA (A + B)^-1 B + colorredB (A + B)^-1 B &=& (A + B)(A + B)^-1 B\
&=& B\
&=& B (A + B)^-1 (A + B) \
&=& colorblueB (A + B)^-1 A + colorredB (A + B)^-1 B\
endeqnarray
$$
because matrix addition is commutative. The identity follows by cancelling $B (A + B)^-1 B$. (If there is a simpler linear proof, say, without needing cancellation, please say so. That extra weird matrix, while obviously looking right and obviously doing the job, just pops out of nowhere. Or does it?)
Algebra is blind manipulation of symbols. The identity holds in any abstract ring with multiplicative inverses. But for a given model, the identity says something about that model. I've only seen the identity in the context of matrices, but I don't see what's so special about it there.
What does this identity do for matrices?
Is there something, in matrix theory, for which this is special? Does it ever really come up in proofs? $(A+B)^-1$ can't be the only matrices that allow such quasi-commutativity, can they? Is there a visualization or geometric interpretation or a meaningful anything interpretation?
intuition matrix-equations geometric-interpretation
1
the title is not same as question?
â mathreadler
Jul 14 at 22:18
Yes it looks kind of curious indeed.
â mathreadler
Jul 14 at 22:23
It seems to be same pseudo commutativity or what to call it also if instead $(A-B)^-1$ in middle. But then both reduce to 0 matrix instead of B in middle if doing same thing for proof.
â mathreadler
Jul 14 at 22:35
2
Any linear combination $(rA+sB)^-1$ will work for non-zero numbers $r,s$. Just replace $A$ with $rA$ and $B$ with $sB$ in the proof and divide by $rs$ at the final step.
â John Wayland Bales
Jul 14 at 22:44
So this implies that it is true for any ring when $a+b$ is invertible.
â Thomas Andrews
Jul 14 at 22:44
 |Â
show 5 more comments
up vote
5
down vote
favorite
up vote
5
down vote
favorite
I am curious about a seemingly simple identity in matrix algebra. Though matrix multiplication is not commutative (the classic example of noncommutativity, it does allow a commutativity of sorts around a very specific third matrix:
$$
colorblueA (A + B)^-1 B = colorblueB (A + B)^-1 A
$$
There is a very simple algebraic proof:
$$
begineqnarray
colorblueA (A + B)^-1 B + colorredB (A + B)^-1 B &=& (A + B)(A + B)^-1 B\
&=& B\
&=& B (A + B)^-1 (A + B) \
&=& colorblueB (A + B)^-1 A + colorredB (A + B)^-1 B\
endeqnarray
$$
because matrix addition is commutative. The identity follows by cancelling $B (A + B)^-1 B$. (If there is a simpler linear proof, say, without needing cancellation, please say so. That extra weird matrix, while obviously looking right and obviously doing the job, just pops out of nowhere. Or does it?)
Algebra is blind manipulation of symbols. The identity holds in any abstract ring with multiplicative inverses. But for a given model, the identity says something about that model. I've only seen the identity in the context of matrices, but I don't see what's so special about it there.
What does this identity do for matrices?
Is there something, in matrix theory, for which this is special? Does it ever really come up in proofs? $(A+B)^-1$ can't be the only matrices that allow such quasi-commutativity, can they? Is there a visualization or geometric interpretation or a meaningful anything interpretation?
intuition matrix-equations geometric-interpretation
I am curious about a seemingly simple identity in matrix algebra. Though matrix multiplication is not commutative (the classic example of noncommutativity, it does allow a commutativity of sorts around a very specific third matrix:
$$
colorblueA (A + B)^-1 B = colorblueB (A + B)^-1 A
$$
There is a very simple algebraic proof:
$$
begineqnarray
colorblueA (A + B)^-1 B + colorredB (A + B)^-1 B &=& (A + B)(A + B)^-1 B\
&=& B\
&=& B (A + B)^-1 (A + B) \
&=& colorblueB (A + B)^-1 A + colorredB (A + B)^-1 B\
endeqnarray
$$
because matrix addition is commutative. The identity follows by cancelling $B (A + B)^-1 B$. (If there is a simpler linear proof, say, without needing cancellation, please say so. That extra weird matrix, while obviously looking right and obviously doing the job, just pops out of nowhere. Or does it?)
Algebra is blind manipulation of symbols. The identity holds in any abstract ring with multiplicative inverses. But for a given model, the identity says something about that model. I've only seen the identity in the context of matrices, but I don't see what's so special about it there.
What does this identity do for matrices?
Is there something, in matrix theory, for which this is special? Does it ever really come up in proofs? $(A+B)^-1$ can't be the only matrices that allow such quasi-commutativity, can they? Is there a visualization or geometric interpretation or a meaningful anything interpretation?
intuition matrix-equations geometric-interpretation
edited Jul 17 at 12:37
asked Jul 14 at 22:11
Mitch
5,9112456
5,9112456
1
the title is not same as question?
â mathreadler
Jul 14 at 22:18
Yes it looks kind of curious indeed.
â mathreadler
Jul 14 at 22:23
It seems to be same pseudo commutativity or what to call it also if instead $(A-B)^-1$ in middle. But then both reduce to 0 matrix instead of B in middle if doing same thing for proof.
â mathreadler
Jul 14 at 22:35
2
Any linear combination $(rA+sB)^-1$ will work for non-zero numbers $r,s$. Just replace $A$ with $rA$ and $B$ with $sB$ in the proof and divide by $rs$ at the final step.
â John Wayland Bales
Jul 14 at 22:44
So this implies that it is true for any ring when $a+b$ is invertible.
â Thomas Andrews
Jul 14 at 22:44
 |Â
show 5 more comments
1
the title is not same as question?
â mathreadler
Jul 14 at 22:18
Yes it looks kind of curious indeed.
â mathreadler
Jul 14 at 22:23
It seems to be same pseudo commutativity or what to call it also if instead $(A-B)^-1$ in middle. But then both reduce to 0 matrix instead of B in middle if doing same thing for proof.
â mathreadler
Jul 14 at 22:35
2
Any linear combination $(rA+sB)^-1$ will work for non-zero numbers $r,s$. Just replace $A$ with $rA$ and $B$ with $sB$ in the proof and divide by $rs$ at the final step.
â John Wayland Bales
Jul 14 at 22:44
So this implies that it is true for any ring when $a+b$ is invertible.
â Thomas Andrews
Jul 14 at 22:44
1
1
the title is not same as question?
â mathreadler
Jul 14 at 22:18
the title is not same as question?
â mathreadler
Jul 14 at 22:18
Yes it looks kind of curious indeed.
â mathreadler
Jul 14 at 22:23
Yes it looks kind of curious indeed.
â mathreadler
Jul 14 at 22:23
It seems to be same pseudo commutativity or what to call it also if instead $(A-B)^-1$ in middle. But then both reduce to 0 matrix instead of B in middle if doing same thing for proof.
â mathreadler
Jul 14 at 22:35
It seems to be same pseudo commutativity or what to call it also if instead $(A-B)^-1$ in middle. But then both reduce to 0 matrix instead of B in middle if doing same thing for proof.
â mathreadler
Jul 14 at 22:35
2
2
Any linear combination $(rA+sB)^-1$ will work for non-zero numbers $r,s$. Just replace $A$ with $rA$ and $B$ with $sB$ in the proof and divide by $rs$ at the final step.
â John Wayland Bales
Jul 14 at 22:44
Any linear combination $(rA+sB)^-1$ will work for non-zero numbers $r,s$. Just replace $A$ with $rA$ and $B$ with $sB$ in the proof and divide by $rs$ at the final step.
â John Wayland Bales
Jul 14 at 22:44
So this implies that it is true for any ring when $a+b$ is invertible.
â Thomas Andrews
Jul 14 at 22:44
So this implies that it is true for any ring when $a+b$ is invertible.
â Thomas Andrews
Jul 14 at 22:44
 |Â
show 5 more comments
3 Answers
3
active
oldest
votes
up vote
7
down vote
In the cases when $A,B$ and $A+B$ are all invertible, the inverse of the equality is:
$$B^-1(A+B)A^-1=A^-1(A+B)B^-1$$
But simple calculation shows the left side is $B^-1+A^-1,$ and the right side is $A^-1+B^-1.$
The general case can be shown by noticing that the set of pairs $(A,B)$ such that $A,B,A+B$ are invertible is dense in the set of pairs $A,B$ such that $A+B$ is invertible. Since the function is continous, that would finish it.
We can do this by taking an arbitrary $A,B$ and then replacing it with $A+lambda I, B-lambda I,$ where $lambda$ is a positive value with smaller magnitude than any of the non-zero eigenvalues of $A,B.$
That continuity argument, of course, doesn't extend to matrices over discrete fields, and this equality is true in any ring. If $R$ is a ring (with identity $I$) and $a,c,din R$ so that $dc=cd=I,(*)$ then:
$$ad(c-a)=(c-a)da,tag1$$
because he left side is $adc-ada=a-ada$ and the right side is $cda-ada=a-ada,$ so they are equal.
Now given $a,bin R$ so that $a+b$ is invertible in $R,$ let $c=a+b,d=(a+b)^-1$. Then $b=c-a$ so (1) becomes $$a(a+b)^-1b=b(a+b)^-1a$$
(*) There are rings where $cd=I$ does not imply $dc=I,$ but in square matrices, $DC=I $ means $CD=I.$ So for the general ring, we need the condition $cd=I$ and $dc=I.$
I'd be interested in seeing the details of the density result you mention!
â Ryan Gibara
Jul 15 at 16:59
Fix $C$ invertible, say $CD=I$ Consider the set $X$ of all pairs of matrices $(A,B)$ such that $A+B=C.$ Then $f_C(A,B)=ADB-BDA$ is a continuous function $Xto M_n.$ In the subset $Ysubseteq X$ of pairs $(A,B)$ with $A,B$ invertible, t $f(A,B)=0.$ Now, given any $(A,B)in X$, and $epsilon>0$, there is a positive $lambda<epsilon$ such that $A'=A-lambda I$ and $B+lambda I$ are invertible. So we can find a pair $(A',B')in Y$ arbitrarily close to $(A,B)in X,$ and since $f(A',B')=0,$ for some point arbitrarily close to $(A,B),$ we must have $f(A,B)=0$ too. @RyanGibara
â Thomas Andrews
Jul 15 at 17:27
@RyanGibara the details actually weren't important for the "intuition" part. The theorem is just much more obvious is $A,B,$ and $A+B$ are all invertible, and the intuition is that invertible matrices are not just dense in the set of all matrices but, in some sense, "very dense." In a neighborhood of a non-invertible matrix $A$, the non-invertible matrices have measure zero. Given the set $Q$ of matrices such that $|Q|<epsilon$, just picking a random $Q$, the probability that $A+Q$ and $B-Q$ is invertible is $1.$
â Thomas Andrews
Jul 15 at 17:46
add a comment |Â
up vote
2
down vote
The only thing it reminds me of right away is $$fracaba+b = left(fraca+babright)^-1 = left(frac 1 a + frac 1 bright)^-1$$
Which is "harmonic sum". Related to harmonic mean for scalars $a,b$.
In other words harmonic means are well defined without respect to order for matrices(?)
If three terms $$left(frac 1 a + frac 1 b + frac 1 cright)^-1 = left(frac bc+ac+ab abcright)^-1 = frac abcbc+ac+ab = frac a(bc)bc+a(c+b)$$
I wonder if something can be proven for $$A(BC+A(B+C))^-1(BC)=(BC)(BC+A(B+C))^-1(A)$$
add a comment |Â
up vote
0
down vote
Just put
$$
A = left( A + B right) - B
$$
to get
$$
eqalign
& left( left( A + B right) - B right)left( A + B right)^, - ,1 B = Bleft( A + B right)^, - ,1 left( left( A + B right) - B right) cr
& quad Downarrow cr
& B - Bleft( A + B right)^, - ,1 B = B - Bleft( A + B right)^, - ,1 B cr
$$
Maybe a better insight on what is going on can be obtained by putting
$$
left{ matrix
A = 1 over 2left( A + B right) - 1 over 2left( B - A right) hfill cr
B = 1 over 2left( A + B right) + 1 over 2left( B - A right) hfill cr right.
$$
which then gives
$$
eqalign
& 1 over 4left( left( A + B right) - left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) + left( B - A right) right) = cr
& = 1 over 4left( left( A + B right) + left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) - left( B - A right) right) = cr
& quad Downarrow cr
& left( I - left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) + left( B - A right) right) = cr
& = left( I + left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) - left( B - A right) right) cr
& quad Downarrow cr
& left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) = cr
& = left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) cr
$$
add a comment |Â
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
7
down vote
In the cases when $A,B$ and $A+B$ are all invertible, the inverse of the equality is:
$$B^-1(A+B)A^-1=A^-1(A+B)B^-1$$
But simple calculation shows the left side is $B^-1+A^-1,$ and the right side is $A^-1+B^-1.$
The general case can be shown by noticing that the set of pairs $(A,B)$ such that $A,B,A+B$ are invertible is dense in the set of pairs $A,B$ such that $A+B$ is invertible. Since the function is continous, that would finish it.
We can do this by taking an arbitrary $A,B$ and then replacing it with $A+lambda I, B-lambda I,$ where $lambda$ is a positive value with smaller magnitude than any of the non-zero eigenvalues of $A,B.$
That continuity argument, of course, doesn't extend to matrices over discrete fields, and this equality is true in any ring. If $R$ is a ring (with identity $I$) and $a,c,din R$ so that $dc=cd=I,(*)$ then:
$$ad(c-a)=(c-a)da,tag1$$
because he left side is $adc-ada=a-ada$ and the right side is $cda-ada=a-ada,$ so they are equal.
Now given $a,bin R$ so that $a+b$ is invertible in $R,$ let $c=a+b,d=(a+b)^-1$. Then $b=c-a$ so (1) becomes $$a(a+b)^-1b=b(a+b)^-1a$$
(*) There are rings where $cd=I$ does not imply $dc=I,$ but in square matrices, $DC=I $ means $CD=I.$ So for the general ring, we need the condition $cd=I$ and $dc=I.$
I'd be interested in seeing the details of the density result you mention!
â Ryan Gibara
Jul 15 at 16:59
Fix $C$ invertible, say $CD=I$ Consider the set $X$ of all pairs of matrices $(A,B)$ such that $A+B=C.$ Then $f_C(A,B)=ADB-BDA$ is a continuous function $Xto M_n.$ In the subset $Ysubseteq X$ of pairs $(A,B)$ with $A,B$ invertible, t $f(A,B)=0.$ Now, given any $(A,B)in X$, and $epsilon>0$, there is a positive $lambda<epsilon$ such that $A'=A-lambda I$ and $B+lambda I$ are invertible. So we can find a pair $(A',B')in Y$ arbitrarily close to $(A,B)in X,$ and since $f(A',B')=0,$ for some point arbitrarily close to $(A,B),$ we must have $f(A,B)=0$ too. @RyanGibara
â Thomas Andrews
Jul 15 at 17:27
@RyanGibara the details actually weren't important for the "intuition" part. The theorem is just much more obvious is $A,B,$ and $A+B$ are all invertible, and the intuition is that invertible matrices are not just dense in the set of all matrices but, in some sense, "very dense." In a neighborhood of a non-invertible matrix $A$, the non-invertible matrices have measure zero. Given the set $Q$ of matrices such that $|Q|<epsilon$, just picking a random $Q$, the probability that $A+Q$ and $B-Q$ is invertible is $1.$
â Thomas Andrews
Jul 15 at 17:46
add a comment |Â
up vote
7
down vote
In the cases when $A,B$ and $A+B$ are all invertible, the inverse of the equality is:
$$B^-1(A+B)A^-1=A^-1(A+B)B^-1$$
But simple calculation shows the left side is $B^-1+A^-1,$ and the right side is $A^-1+B^-1.$
The general case can be shown by noticing that the set of pairs $(A,B)$ such that $A,B,A+B$ are invertible is dense in the set of pairs $A,B$ such that $A+B$ is invertible. Since the function is continous, that would finish it.
We can do this by taking an arbitrary $A,B$ and then replacing it with $A+lambda I, B-lambda I,$ where $lambda$ is a positive value with smaller magnitude than any of the non-zero eigenvalues of $A,B.$
That continuity argument, of course, doesn't extend to matrices over discrete fields, and this equality is true in any ring. If $R$ is a ring (with identity $I$) and $a,c,din R$ so that $dc=cd=I,(*)$ then:
$$ad(c-a)=(c-a)da,tag1$$
because he left side is $adc-ada=a-ada$ and the right side is $cda-ada=a-ada,$ so they are equal.
Now given $a,bin R$ so that $a+b$ is invertible in $R,$ let $c=a+b,d=(a+b)^-1$. Then $b=c-a$ so (1) becomes $$a(a+b)^-1b=b(a+b)^-1a$$
(*) There are rings where $cd=I$ does not imply $dc=I,$ but in square matrices, $DC=I $ means $CD=I.$ So for the general ring, we need the condition $cd=I$ and $dc=I.$
I'd be interested in seeing the details of the density result you mention!
â Ryan Gibara
Jul 15 at 16:59
Fix $C$ invertible, say $CD=I$ Consider the set $X$ of all pairs of matrices $(A,B)$ such that $A+B=C.$ Then $f_C(A,B)=ADB-BDA$ is a continuous function $Xto M_n.$ In the subset $Ysubseteq X$ of pairs $(A,B)$ with $A,B$ invertible, t $f(A,B)=0.$ Now, given any $(A,B)in X$, and $epsilon>0$, there is a positive $lambda<epsilon$ such that $A'=A-lambda I$ and $B+lambda I$ are invertible. So we can find a pair $(A',B')in Y$ arbitrarily close to $(A,B)in X,$ and since $f(A',B')=0,$ for some point arbitrarily close to $(A,B),$ we must have $f(A,B)=0$ too. @RyanGibara
â Thomas Andrews
Jul 15 at 17:27
@RyanGibara the details actually weren't important for the "intuition" part. The theorem is just much more obvious is $A,B,$ and $A+B$ are all invertible, and the intuition is that invertible matrices are not just dense in the set of all matrices but, in some sense, "very dense." In a neighborhood of a non-invertible matrix $A$, the non-invertible matrices have measure zero. Given the set $Q$ of matrices such that $|Q|<epsilon$, just picking a random $Q$, the probability that $A+Q$ and $B-Q$ is invertible is $1.$
â Thomas Andrews
Jul 15 at 17:46
add a comment |Â
up vote
7
down vote
up vote
7
down vote
In the cases when $A,B$ and $A+B$ are all invertible, the inverse of the equality is:
$$B^-1(A+B)A^-1=A^-1(A+B)B^-1$$
But simple calculation shows the left side is $B^-1+A^-1,$ and the right side is $A^-1+B^-1.$
The general case can be shown by noticing that the set of pairs $(A,B)$ such that $A,B,A+B$ are invertible is dense in the set of pairs $A,B$ such that $A+B$ is invertible. Since the function is continous, that would finish it.
We can do this by taking an arbitrary $A,B$ and then replacing it with $A+lambda I, B-lambda I,$ where $lambda$ is a positive value with smaller magnitude than any of the non-zero eigenvalues of $A,B.$
That continuity argument, of course, doesn't extend to matrices over discrete fields, and this equality is true in any ring. If $R$ is a ring (with identity $I$) and $a,c,din R$ so that $dc=cd=I,(*)$ then:
$$ad(c-a)=(c-a)da,tag1$$
because he left side is $adc-ada=a-ada$ and the right side is $cda-ada=a-ada,$ so they are equal.
Now given $a,bin R$ so that $a+b$ is invertible in $R,$ let $c=a+b,d=(a+b)^-1$. Then $b=c-a$ so (1) becomes $$a(a+b)^-1b=b(a+b)^-1a$$
(*) There are rings where $cd=I$ does not imply $dc=I,$ but in square matrices, $DC=I $ means $CD=I.$ So for the general ring, we need the condition $cd=I$ and $dc=I.$
In the cases when $A,B$ and $A+B$ are all invertible, the inverse of the equality is:
$$B^-1(A+B)A^-1=A^-1(A+B)B^-1$$
But simple calculation shows the left side is $B^-1+A^-1,$ and the right side is $A^-1+B^-1.$
The general case can be shown by noticing that the set of pairs $(A,B)$ such that $A,B,A+B$ are invertible is dense in the set of pairs $A,B$ such that $A+B$ is invertible. Since the function is continous, that would finish it.
We can do this by taking an arbitrary $A,B$ and then replacing it with $A+lambda I, B-lambda I,$ where $lambda$ is a positive value with smaller magnitude than any of the non-zero eigenvalues of $A,B.$
That continuity argument, of course, doesn't extend to matrices over discrete fields, and this equality is true in any ring. If $R$ is a ring (with identity $I$) and $a,c,din R$ so that $dc=cd=I,(*)$ then:
$$ad(c-a)=(c-a)da,tag1$$
because he left side is $adc-ada=a-ada$ and the right side is $cda-ada=a-ada,$ so they are equal.
Now given $a,bin R$ so that $a+b$ is invertible in $R,$ let $c=a+b,d=(a+b)^-1$. Then $b=c-a$ so (1) becomes $$a(a+b)^-1b=b(a+b)^-1a$$
(*) There are rings where $cd=I$ does not imply $dc=I,$ but in square matrices, $DC=I $ means $CD=I.$ So for the general ring, we need the condition $cd=I$ and $dc=I.$
edited Jul 15 at 16:31
answered Jul 14 at 23:05
Thomas Andrews
128k10144285
128k10144285
I'd be interested in seeing the details of the density result you mention!
â Ryan Gibara
Jul 15 at 16:59
Fix $C$ invertible, say $CD=I$ Consider the set $X$ of all pairs of matrices $(A,B)$ such that $A+B=C.$ Then $f_C(A,B)=ADB-BDA$ is a continuous function $Xto M_n.$ In the subset $Ysubseteq X$ of pairs $(A,B)$ with $A,B$ invertible, t $f(A,B)=0.$ Now, given any $(A,B)in X$, and $epsilon>0$, there is a positive $lambda<epsilon$ such that $A'=A-lambda I$ and $B+lambda I$ are invertible. So we can find a pair $(A',B')in Y$ arbitrarily close to $(A,B)in X,$ and since $f(A',B')=0,$ for some point arbitrarily close to $(A,B),$ we must have $f(A,B)=0$ too. @RyanGibara
â Thomas Andrews
Jul 15 at 17:27
@RyanGibara the details actually weren't important for the "intuition" part. The theorem is just much more obvious is $A,B,$ and $A+B$ are all invertible, and the intuition is that invertible matrices are not just dense in the set of all matrices but, in some sense, "very dense." In a neighborhood of a non-invertible matrix $A$, the non-invertible matrices have measure zero. Given the set $Q$ of matrices such that $|Q|<epsilon$, just picking a random $Q$, the probability that $A+Q$ and $B-Q$ is invertible is $1.$
â Thomas Andrews
Jul 15 at 17:46
add a comment |Â
I'd be interested in seeing the details of the density result you mention!
â Ryan Gibara
Jul 15 at 16:59
Fix $C$ invertible, say $CD=I$ Consider the set $X$ of all pairs of matrices $(A,B)$ such that $A+B=C.$ Then $f_C(A,B)=ADB-BDA$ is a continuous function $Xto M_n.$ In the subset $Ysubseteq X$ of pairs $(A,B)$ with $A,B$ invertible, t $f(A,B)=0.$ Now, given any $(A,B)in X$, and $epsilon>0$, there is a positive $lambda<epsilon$ such that $A'=A-lambda I$ and $B+lambda I$ are invertible. So we can find a pair $(A',B')in Y$ arbitrarily close to $(A,B)in X,$ and since $f(A',B')=0,$ for some point arbitrarily close to $(A,B),$ we must have $f(A,B)=0$ too. @RyanGibara
â Thomas Andrews
Jul 15 at 17:27
@RyanGibara the details actually weren't important for the "intuition" part. The theorem is just much more obvious is $A,B,$ and $A+B$ are all invertible, and the intuition is that invertible matrices are not just dense in the set of all matrices but, in some sense, "very dense." In a neighborhood of a non-invertible matrix $A$, the non-invertible matrices have measure zero. Given the set $Q$ of matrices such that $|Q|<epsilon$, just picking a random $Q$, the probability that $A+Q$ and $B-Q$ is invertible is $1.$
â Thomas Andrews
Jul 15 at 17:46
I'd be interested in seeing the details of the density result you mention!
â Ryan Gibara
Jul 15 at 16:59
I'd be interested in seeing the details of the density result you mention!
â Ryan Gibara
Jul 15 at 16:59
Fix $C$ invertible, say $CD=I$ Consider the set $X$ of all pairs of matrices $(A,B)$ such that $A+B=C.$ Then $f_C(A,B)=ADB-BDA$ is a continuous function $Xto M_n.$ In the subset $Ysubseteq X$ of pairs $(A,B)$ with $A,B$ invertible, t $f(A,B)=0.$ Now, given any $(A,B)in X$, and $epsilon>0$, there is a positive $lambda<epsilon$ such that $A'=A-lambda I$ and $B+lambda I$ are invertible. So we can find a pair $(A',B')in Y$ arbitrarily close to $(A,B)in X,$ and since $f(A',B')=0,$ for some point arbitrarily close to $(A,B),$ we must have $f(A,B)=0$ too. @RyanGibara
â Thomas Andrews
Jul 15 at 17:27
Fix $C$ invertible, say $CD=I$ Consider the set $X$ of all pairs of matrices $(A,B)$ such that $A+B=C.$ Then $f_C(A,B)=ADB-BDA$ is a continuous function $Xto M_n.$ In the subset $Ysubseteq X$ of pairs $(A,B)$ with $A,B$ invertible, t $f(A,B)=0.$ Now, given any $(A,B)in X$, and $epsilon>0$, there is a positive $lambda<epsilon$ such that $A'=A-lambda I$ and $B+lambda I$ are invertible. So we can find a pair $(A',B')in Y$ arbitrarily close to $(A,B)in X,$ and since $f(A',B')=0,$ for some point arbitrarily close to $(A,B),$ we must have $f(A,B)=0$ too. @RyanGibara
â Thomas Andrews
Jul 15 at 17:27
@RyanGibara the details actually weren't important for the "intuition" part. The theorem is just much more obvious is $A,B,$ and $A+B$ are all invertible, and the intuition is that invertible matrices are not just dense in the set of all matrices but, in some sense, "very dense." In a neighborhood of a non-invertible matrix $A$, the non-invertible matrices have measure zero. Given the set $Q$ of matrices such that $|Q|<epsilon$, just picking a random $Q$, the probability that $A+Q$ and $B-Q$ is invertible is $1.$
â Thomas Andrews
Jul 15 at 17:46
@RyanGibara the details actually weren't important for the "intuition" part. The theorem is just much more obvious is $A,B,$ and $A+B$ are all invertible, and the intuition is that invertible matrices are not just dense in the set of all matrices but, in some sense, "very dense." In a neighborhood of a non-invertible matrix $A$, the non-invertible matrices have measure zero. Given the set $Q$ of matrices such that $|Q|<epsilon$, just picking a random $Q$, the probability that $A+Q$ and $B-Q$ is invertible is $1.$
â Thomas Andrews
Jul 15 at 17:46
add a comment |Â
up vote
2
down vote
The only thing it reminds me of right away is $$fracaba+b = left(fraca+babright)^-1 = left(frac 1 a + frac 1 bright)^-1$$
Which is "harmonic sum". Related to harmonic mean for scalars $a,b$.
In other words harmonic means are well defined without respect to order for matrices(?)
If three terms $$left(frac 1 a + frac 1 b + frac 1 cright)^-1 = left(frac bc+ac+ab abcright)^-1 = frac abcbc+ac+ab = frac a(bc)bc+a(c+b)$$
I wonder if something can be proven for $$A(BC+A(B+C))^-1(BC)=(BC)(BC+A(B+C))^-1(A)$$
add a comment |Â
up vote
2
down vote
The only thing it reminds me of right away is $$fracaba+b = left(fraca+babright)^-1 = left(frac 1 a + frac 1 bright)^-1$$
Which is "harmonic sum". Related to harmonic mean for scalars $a,b$.
In other words harmonic means are well defined without respect to order for matrices(?)
If three terms $$left(frac 1 a + frac 1 b + frac 1 cright)^-1 = left(frac bc+ac+ab abcright)^-1 = frac abcbc+ac+ab = frac a(bc)bc+a(c+b)$$
I wonder if something can be proven for $$A(BC+A(B+C))^-1(BC)=(BC)(BC+A(B+C))^-1(A)$$
add a comment |Â
up vote
2
down vote
up vote
2
down vote
The only thing it reminds me of right away is $$fracaba+b = left(fraca+babright)^-1 = left(frac 1 a + frac 1 bright)^-1$$
Which is "harmonic sum". Related to harmonic mean for scalars $a,b$.
In other words harmonic means are well defined without respect to order for matrices(?)
If three terms $$left(frac 1 a + frac 1 b + frac 1 cright)^-1 = left(frac bc+ac+ab abcright)^-1 = frac abcbc+ac+ab = frac a(bc)bc+a(c+b)$$
I wonder if something can be proven for $$A(BC+A(B+C))^-1(BC)=(BC)(BC+A(B+C))^-1(A)$$
The only thing it reminds me of right away is $$fracaba+b = left(fraca+babright)^-1 = left(frac 1 a + frac 1 bright)^-1$$
Which is "harmonic sum". Related to harmonic mean for scalars $a,b$.
In other words harmonic means are well defined without respect to order for matrices(?)
If three terms $$left(frac 1 a + frac 1 b + frac 1 cright)^-1 = left(frac bc+ac+ab abcright)^-1 = frac abcbc+ac+ab = frac a(bc)bc+a(c+b)$$
I wonder if something can be proven for $$A(BC+A(B+C))^-1(BC)=(BC)(BC+A(B+C))^-1(A)$$
edited Jul 14 at 23:34
answered Jul 14 at 23:04
mathreadler
13.6k71857
13.6k71857
add a comment |Â
add a comment |Â
up vote
0
down vote
Just put
$$
A = left( A + B right) - B
$$
to get
$$
eqalign
& left( left( A + B right) - B right)left( A + B right)^, - ,1 B = Bleft( A + B right)^, - ,1 left( left( A + B right) - B right) cr
& quad Downarrow cr
& B - Bleft( A + B right)^, - ,1 B = B - Bleft( A + B right)^, - ,1 B cr
$$
Maybe a better insight on what is going on can be obtained by putting
$$
left{ matrix
A = 1 over 2left( A + B right) - 1 over 2left( B - A right) hfill cr
B = 1 over 2left( A + B right) + 1 over 2left( B - A right) hfill cr right.
$$
which then gives
$$
eqalign
& 1 over 4left( left( A + B right) - left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) + left( B - A right) right) = cr
& = 1 over 4left( left( A + B right) + left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) - left( B - A right) right) = cr
& quad Downarrow cr
& left( I - left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) + left( B - A right) right) = cr
& = left( I + left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) - left( B - A right) right) cr
& quad Downarrow cr
& left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) = cr
& = left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) cr
$$
add a comment |Â
up vote
0
down vote
Just put
$$
A = left( A + B right) - B
$$
to get
$$
eqalign
& left( left( A + B right) - B right)left( A + B right)^, - ,1 B = Bleft( A + B right)^, - ,1 left( left( A + B right) - B right) cr
& quad Downarrow cr
& B - Bleft( A + B right)^, - ,1 B = B - Bleft( A + B right)^, - ,1 B cr
$$
Maybe a better insight on what is going on can be obtained by putting
$$
left{ matrix
A = 1 over 2left( A + B right) - 1 over 2left( B - A right) hfill cr
B = 1 over 2left( A + B right) + 1 over 2left( B - A right) hfill cr right.
$$
which then gives
$$
eqalign
& 1 over 4left( left( A + B right) - left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) + left( B - A right) right) = cr
& = 1 over 4left( left( A + B right) + left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) - left( B - A right) right) = cr
& quad Downarrow cr
& left( I - left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) + left( B - A right) right) = cr
& = left( I + left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) - left( B - A right) right) cr
& quad Downarrow cr
& left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) = cr
& = left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) cr
$$
add a comment |Â
up vote
0
down vote
up vote
0
down vote
Just put
$$
A = left( A + B right) - B
$$
to get
$$
eqalign
& left( left( A + B right) - B right)left( A + B right)^, - ,1 B = Bleft( A + B right)^, - ,1 left( left( A + B right) - B right) cr
& quad Downarrow cr
& B - Bleft( A + B right)^, - ,1 B = B - Bleft( A + B right)^, - ,1 B cr
$$
Maybe a better insight on what is going on can be obtained by putting
$$
left{ matrix
A = 1 over 2left( A + B right) - 1 over 2left( B - A right) hfill cr
B = 1 over 2left( A + B right) + 1 over 2left( B - A right) hfill cr right.
$$
which then gives
$$
eqalign
& 1 over 4left( left( A + B right) - left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) + left( B - A right) right) = cr
& = 1 over 4left( left( A + B right) + left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) - left( B - A right) right) = cr
& quad Downarrow cr
& left( I - left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) + left( B - A right) right) = cr
& = left( I + left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) - left( B - A right) right) cr
& quad Downarrow cr
& left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) = cr
& = left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) cr
$$
Just put
$$
A = left( A + B right) - B
$$
to get
$$
eqalign
& left( left( A + B right) - B right)left( A + B right)^, - ,1 B = Bleft( A + B right)^, - ,1 left( left( A + B right) - B right) cr
& quad Downarrow cr
& B - Bleft( A + B right)^, - ,1 B = B - Bleft( A + B right)^, - ,1 B cr
$$
Maybe a better insight on what is going on can be obtained by putting
$$
left{ matrix
A = 1 over 2left( A + B right) - 1 over 2left( B - A right) hfill cr
B = 1 over 2left( A + B right) + 1 over 2left( B - A right) hfill cr right.
$$
which then gives
$$
eqalign
& 1 over 4left( left( A + B right) - left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) + left( B - A right) right) = cr
& = 1 over 4left( left( A + B right) + left( B - A right) right)left( A + B right)^, - ,1 left( left( A + B right) - left( B - A right) right) = cr
& quad Downarrow cr
& left( I - left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) + left( B - A right) right) = cr
& = left( I + left( B - A right)left( A + B right)^, - ,1 right)left( left( A + B right) - left( B - A right) right) cr
& quad Downarrow cr
& left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) = cr
& = left( A + B right) - left( B - A right)left( A + B right)^, - ,1 left( B - A right) cr
$$
edited Jul 14 at 23:57
answered Jul 14 at 23:42
G Cab
15.1k31136
15.1k31136
add a comment |Â
add a comment |Â
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2852002%2fgeometric-or-matrix-intuition-on-aa-b-1b-b-a-b-1-a%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
1
the title is not same as question?
â mathreadler
Jul 14 at 22:18
Yes it looks kind of curious indeed.
â mathreadler
Jul 14 at 22:23
It seems to be same pseudo commutativity or what to call it also if instead $(A-B)^-1$ in middle. But then both reduce to 0 matrix instead of B in middle if doing same thing for proof.
â mathreadler
Jul 14 at 22:35
2
Any linear combination $(rA+sB)^-1$ will work for non-zero numbers $r,s$. Just replace $A$ with $rA$ and $B$ with $sB$ in the proof and divide by $rs$ at the final step.
â John Wayland Bales
Jul 14 at 22:44
So this implies that it is true for any ring when $a+b$ is invertible.
â Thomas Andrews
Jul 14 at 22:44