Geometric Interpretation of Determinant of Transpose
$begingroup$
Below are two well-known statements regarding the determinant function:
- When $A$ is a square matrix, $det(A)$ is the signed volume of the parallelepiped whose edges are columns of $A$.
- When $A$ is a square matrix, $det(A) = det(A^T)$.
It appears to me that 1 is how $det$ got invented in the first place. The multilinearity and alternating property are natural consequences. Cramer's rule can also be easily understood geometrically. However, 1 does not immediately yield 2.
Of course the proof of 2 can be done algebraically, and it's not wrong to say that the algebra involved does come from 1. My question is, can 2 be viewed more geometrically? Can it feed us back a better geometric interpretation?
[How would one visualize the row vectors of $A$ from the column vectors? And how would one think, geometrically, about the equality of the signed volume spanned by two visually different parallelepipeds?]
Edit: If you are going to use multiplicativity of the determinant, please provide a geometric intuition to that too. Keep reading to see what I mean.
Here is one of my failed attempts. I introduce a different interpretation:
- $det(A)$ is the ratio of the volume of an arbitrary parallelepiped changed by applying $A^T$ (left multiplication), viewed as a linear transformation.
The reason I say applying $A^T$ instead of applying $A$ is that I want the transpose to appear somewhere, and coincidentally the computation of $A^T u$ can be thought of geometrically in terms of columns of $A$: $A^T u$ is the vector whose components are dot products of columns of $A$ with $u$. (I'll take for granted the validity of the word arbitrary in 3 for the moment.) What is left to argue is the equality of 1 and 3, which I have not been able to do geometrically.
Is This a Duplicate Question?
In a sense yes, and in a sense no. I did participate in that question a long time ago. There was no satisfactory answer. The accepted answer does not say what a "really geometric" interpretation of (*) should be. Yes you can always do certain row and column operations to put a matrix into a diagonal form without affecting the determinant (with an exception of a possible sign flip), and by induction, one can prove $det(A) = det(A^T)$. This, however, seems to me more algebraic than geometric even though each step can be thought of as geometric. I am aware that the property of an interpretation being "geometric" or "algebraic" is not a rigorous notion and is quite subjective. Some may say that this is as far as "geometry" can assist us. I just want to know if that is really the case. That is why I tag this as a soft question. I also secretly hope that maybe one such interpretation would serve as a simple explanation of Cauchy-Binet formula.
linear-algebra geometry soft-question determinant exterior-algebra
$endgroup$
add a comment |
$begingroup$
Below are two well-known statements regarding the determinant function:
- When $A$ is a square matrix, $det(A)$ is the signed volume of the parallelepiped whose edges are columns of $A$.
- When $A$ is a square matrix, $det(A) = det(A^T)$.
It appears to me that 1 is how $det$ got invented in the first place. The multilinearity and alternating property are natural consequences. Cramer's rule can also be easily understood geometrically. However, 1 does not immediately yield 2.
Of course the proof of 2 can be done algebraically, and it's not wrong to say that the algebra involved does come from 1. My question is, can 2 be viewed more geometrically? Can it feed us back a better geometric interpretation?
[How would one visualize the row vectors of $A$ from the column vectors? And how would one think, geometrically, about the equality of the signed volume spanned by two visually different parallelepipeds?]
Edit: If you are going to use multiplicativity of the determinant, please provide a geometric intuition to that too. Keep reading to see what I mean.
Here is one of my failed attempts. I introduce a different interpretation:
- $det(A)$ is the ratio of the volume of an arbitrary parallelepiped changed by applying $A^T$ (left multiplication), viewed as a linear transformation.
The reason I say applying $A^T$ instead of applying $A$ is that I want the transpose to appear somewhere, and coincidentally the computation of $A^T u$ can be thought of geometrically in terms of columns of $A$: $A^T u$ is the vector whose components are dot products of columns of $A$ with $u$. (I'll take for granted the validity of the word arbitrary in 3 for the moment.) What is left to argue is the equality of 1 and 3, which I have not been able to do geometrically.
Is This a Duplicate Question?
In a sense yes, and in a sense no. I did participate in that question a long time ago. There was no satisfactory answer. The accepted answer does not say what a "really geometric" interpretation of (*) should be. Yes you can always do certain row and column operations to put a matrix into a diagonal form without affecting the determinant (with an exception of a possible sign flip), and by induction, one can prove $det(A) = det(A^T)$. This, however, seems to me more algebraic than geometric even though each step can be thought of as geometric. I am aware that the property of an interpretation being "geometric" or "algebraic" is not a rigorous notion and is quite subjective. Some may say that this is as far as "geometry" can assist us. I just want to know if that is really the case. That is why I tag this as a soft question. I also secretly hope that maybe one such interpretation would serve as a simple explanation of Cauchy-Binet formula.
linear-algebra geometry soft-question determinant exterior-algebra
$endgroup$
$begingroup$
This looks like a duplicate of this question
$endgroup$
– preferred_anon
Jun 10 '15 at 12:17
add a comment |
$begingroup$
Below are two well-known statements regarding the determinant function:
- When $A$ is a square matrix, $det(A)$ is the signed volume of the parallelepiped whose edges are columns of $A$.
- When $A$ is a square matrix, $det(A) = det(A^T)$.
It appears to me that 1 is how $det$ got invented in the first place. The multilinearity and alternating property are natural consequences. Cramer's rule can also be easily understood geometrically. However, 1 does not immediately yield 2.
Of course the proof of 2 can be done algebraically, and it's not wrong to say that the algebra involved does come from 1. My question is, can 2 be viewed more geometrically? Can it feed us back a better geometric interpretation?
[How would one visualize the row vectors of $A$ from the column vectors? And how would one think, geometrically, about the equality of the signed volume spanned by two visually different parallelepipeds?]
Edit: If you are going to use multiplicativity of the determinant, please provide a geometric intuition to that too. Keep reading to see what I mean.
Here is one of my failed attempts. I introduce a different interpretation:
- $det(A)$ is the ratio of the volume of an arbitrary parallelepiped changed by applying $A^T$ (left multiplication), viewed as a linear transformation.
The reason I say applying $A^T$ instead of applying $A$ is that I want the transpose to appear somewhere, and coincidentally the computation of $A^T u$ can be thought of geometrically in terms of columns of $A$: $A^T u$ is the vector whose components are dot products of columns of $A$ with $u$. (I'll take for granted the validity of the word arbitrary in 3 for the moment.) What is left to argue is the equality of 1 and 3, which I have not been able to do geometrically.
Is This a Duplicate Question?
In a sense yes, and in a sense no. I did participate in that question a long time ago. There was no satisfactory answer. The accepted answer does not say what a "really geometric" interpretation of (*) should be. Yes you can always do certain row and column operations to put a matrix into a diagonal form without affecting the determinant (with an exception of a possible sign flip), and by induction, one can prove $det(A) = det(A^T)$. This, however, seems to me more algebraic than geometric even though each step can be thought of as geometric. I am aware that the property of an interpretation being "geometric" or "algebraic" is not a rigorous notion and is quite subjective. Some may say that this is as far as "geometry" can assist us. I just want to know if that is really the case. That is why I tag this as a soft question. I also secretly hope that maybe one such interpretation would serve as a simple explanation of Cauchy-Binet formula.
linear-algebra geometry soft-question determinant exterior-algebra
$endgroup$
Below are two well-known statements regarding the determinant function:
- When $A$ is a square matrix, $det(A)$ is the signed volume of the parallelepiped whose edges are columns of $A$.
- When $A$ is a square matrix, $det(A) = det(A^T)$.
It appears to me that 1 is how $det$ got invented in the first place. The multilinearity and alternating property are natural consequences. Cramer's rule can also be easily understood geometrically. However, 1 does not immediately yield 2.
Of course the proof of 2 can be done algebraically, and it's not wrong to say that the algebra involved does come from 1. My question is, can 2 be viewed more geometrically? Can it feed us back a better geometric interpretation?
[How would one visualize the row vectors of $A$ from the column vectors? And how would one think, geometrically, about the equality of the signed volume spanned by two visually different parallelepipeds?]
Edit: If you are going to use multiplicativity of the determinant, please provide a geometric intuition to that too. Keep reading to see what I mean.
Here is one of my failed attempts. I introduce a different interpretation:
- $det(A)$ is the ratio of the volume of an arbitrary parallelepiped changed by applying $A^T$ (left multiplication), viewed as a linear transformation.
The reason I say applying $A^T$ instead of applying $A$ is that I want the transpose to appear somewhere, and coincidentally the computation of $A^T u$ can be thought of geometrically in terms of columns of $A$: $A^T u$ is the vector whose components are dot products of columns of $A$ with $u$. (I'll take for granted the validity of the word arbitrary in 3 for the moment.) What is left to argue is the equality of 1 and 3, which I have not been able to do geometrically.
Is This a Duplicate Question?
In a sense yes, and in a sense no. I did participate in that question a long time ago. There was no satisfactory answer. The accepted answer does not say what a "really geometric" interpretation of (*) should be. Yes you can always do certain row and column operations to put a matrix into a diagonal form without affecting the determinant (with an exception of a possible sign flip), and by induction, one can prove $det(A) = det(A^T)$. This, however, seems to me more algebraic than geometric even though each step can be thought of as geometric. I am aware that the property of an interpretation being "geometric" or "algebraic" is not a rigorous notion and is quite subjective. Some may say that this is as far as "geometry" can assist us. I just want to know if that is really the case. That is why I tag this as a soft question. I also secretly hope that maybe one such interpretation would serve as a simple explanation of Cauchy-Binet formula.
linear-algebra geometry soft-question determinant exterior-algebra
linear-algebra geometry soft-question determinant exterior-algebra
edited Apr 13 '17 at 12:20
Community♦
1
1
asked Jun 10 '15 at 12:04
TunococTunococ
8,1961931
8,1961931
$begingroup$
This looks like a duplicate of this question
$endgroup$
– preferred_anon
Jun 10 '15 at 12:17
add a comment |
$begingroup$
This looks like a duplicate of this question
$endgroup$
– preferred_anon
Jun 10 '15 at 12:17
$begingroup$
This looks like a duplicate of this question
$endgroup$
– preferred_anon
Jun 10 '15 at 12:17
$begingroup$
This looks like a duplicate of this question
$endgroup$
– preferred_anon
Jun 10 '15 at 12:17
add a comment |
4 Answers
4
active
oldest
votes
$begingroup$
Write $A$ as a product of the form $E_1E_2cdots E_kDF_1F_2cdots F_ell$, where each $E_i$ (resp. $F_j$) is an elementary matrix corresponding to row (resp. column) switching or row (resp. column) addition, and $D$ is a diagonal matrix. You are done if you can explain why
- $det A=det(E_1)cdotsdet(E_k)det(D)det(F_1)cdotsdet(F_ell)$ and
- the determinant of each elementary matrix is equal to the determinant of its transpose.
But these are essentially what motivate the definition of determinant: we want to define the determinant as the volume of the parallelepiped generated by the columns of a matrix. Shears do not change the volume (thus the determinant of an elementary matrix for row/column addition is $1$) and scaling in one direction changes the volume in the same proportion (thus $det D$ is the product of $D$'s diagonal entries). Since we want to define the volume as zero in the degenerate case, it follows that together with the previous two requirements, the volume must be signed and by flipping two columns, the sign flips too (thus the determinant of an elementary matrix for row/column switching is $-1$).
$endgroup$
$begingroup$
The proof by using a sequence of shear transformations is precisely what I mentioned in my question. However, your answer did motivate me to think deeper about this proof: How would you "geometrically" explains that the determinant of each elementary matrix is equal to the determinant of its transpose? In other words, if you view columns of the matrix as vectors, why do row operations (changes of basis) affect the volume measurement the way they do?
$endgroup$
– Tunococ
Sep 23 '15 at 23:16
$begingroup$
@Tunococ Isn't that because the transpose of a shear matrix is also a shear matrix? But just how "geometric" is this is debatable. Anyway, I think the crux of the problem is that we on one hand want to define $A^T$ geometrically as the adjoint of $A$ defined on a dual space, but on the other hand we want to compare the parallelepipeds generated by the columns of $A$ and $A^T$ on the same space. Without doing some algebra, I don't think it's easy to see that the transpose of a shear is a shear.
$endgroup$
– user1551
Sep 24 '15 at 2:16
add a comment |
$begingroup$
To prove that the two (generalized) volumes are equal is going to require some algebra -- I don't see a way around that (because the transpose is defined algebraically). This proof doesn't require the use of matrices or characteristic equations or anything, though. I just use a geometric definition of the determinant and then an algebraic formula relating a linear transformation to its adjoint (transpose).
Consider this as the geometric definition of the determinant. If $I$ is the unit $n$-cube in $Bbb R^n$, then we can define the determinant of a linear transformation $T$ as $$underline T(I)=det(T)I$$ This formula tells us that the transformation $underline T$ just scales the unit $n$-cube by its determinant.
For vectors, you might recognize that the adjoint (transpose), $overline T$, of a linear transformation, $T$, is given by $$T(v)cdot w = vcdot overline T(w)$$ However, $I$ is not really a vector, it's an object called an $n$-blade or $n$-vector. So we have to use the modified formula $$langleunderline T(A)Brangle = langle Aoverline T(B)rangle$$ where $langle Mrangle$ returns the scalar part of the multivector $M$.
So for us, it becomes $langleunderline T(I)Irangle = langle Ioverline T(I)rangle$. On the LHS, we get $$langleunderline T(I)Irangle = langle det(T)I^2rangle = det(T)I^2$$ And on the RHS, we get $$langle Ioverline T(I)rangle = langle det(overline T)I^2rangle = det(overline T)I^2$$ Therefore $$det(T) = det(overline T) square$$
$endgroup$
$begingroup$
Yes, I understand that the actual proof has to be algebraic. I'm asking for a more informal explanation that may be able to convince young students that this equality holds without going through too much algebra. By the way, your answer creates another interesting question: What is the geometric intuition of the adjoint? Maybe we can get a better intuition from the definition of adjoint than from the definition of transpose (as in flipping a matrix).
$endgroup$
– Tunococ
Sep 23 '15 at 0:19
add a comment |
$begingroup$
I think there is a geometrical interpretation but it requires a bit of imagination w.r.t inner product spaces. Some notation to start:
$ f: U rightarrow V $ is a linear map and $ f^{*} : V rightarrow U $ is it's adjoint.
Let $ e_1 ..e_n s $ be an orthonormal basis for $U$. Considering only the non-degenerate case, we can see that $ frac{f(e_i)}{lVert f(e_i) rVert} = e^{'}_i $ form a basis for $V$.
Now here is the trick: define $ e^{'}_1..e^{'}_n $ to be an orthonormal basis for $V$. From the perspective of $U$, $e^{'}_is$ may not be orthonormal, but when computing inside of $V$, we can define it to be.
Physically speaking, the space $V$ appears deformed, but only as seen from $U$ (and vice-versa).
From the definition of the adjoint, for any $u in U$ and $ v in V$, we have
$ f(u)cdot v = f^{*}(v)cdot u $
Note that the L.H.S is being computed in $V$, while the R.H.S is being computed in $U$. This equation also holds for $e_is $ and $e^{'}_js$.
$ f(e_i)cdot e^{'}_j = f^{*}(e^{'}_j)cdot e_i$
But the L.H.S is zero for any $ i neq j $!
$ f^{*}(e^{'}_i)cdot e_i = f(e_i)cdot e^{'}_i = lVert f(e_i) rVert $
$ f^{*}(e^{'}_i)cdot e_j = 0$ $ (i neq j) $
Now we come to the crux of the argument. For this special choice of bases, $ f^{*} $ dilates the orthonormal basis vectors $e^{'}_i$ of $V$ by the same amount as $ f$ dilates the orthonormal basis vectors $ e_i$ of $U$. By keeping the order same, it also preserves the orientation.
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
For this reason, $f^{*}$ maps the unit 'volume' of $V$ to the same volume as $f$ does the unit 'volume' of $U$.
Edit 1:
The above treatment is admittedly not very rigorous, to start with determinants are usually defined for linear operators $ f : V rightarrow V $ whereas I stretched the definition to $ f : U rightarrow V $. But here is an alternate way to see the same thing:
for any $ f : V rightarrow V $, the composite transform $ f^{*}f : V rightarrow V $ is a symmetric linear transform (or Matrix), which has an orthornormal eigenvector basis $ e_1, e_2, .. e_n $ with real eigenvalues $ lambda_1, lambda_1,.. lambda_n$
Now consider the $ f(e_i)'s $:
$ f(e_i)cdot f(e_j) = f^{*}f(e_i)cdot e_j = lambda_i e_icdot e_j $
Therefore, $ f(e_i)s $ also form an orthornormal basis like before (except this time not by fiat). Also:
$ f(e_i).f(e_i) = lVert f(e_i) rVert^2 = lambda_i $
Which is to say $f$ scales $ e_i$ by $ sqrt{lambda_i}$. Now set the basis $ e^{'}_i = frac{f(e_i)}{sqrt{lambda_i}} $ for $ f^{*} $ and check the scaling:
$ f^{*}(e^{'}_i) = frac{f^{*}f(e_i)}{sqrt{lambda_i}} = sqrt{lambda_i}e_i$
So again,
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
Edit 2:
Another way to look at this is through the geometrical interpretation of the adjoint itself.
Every linear map, when seen in a certain special basis (as shown above), can be thought of as a composition of two maps: A positive semidefinite map ($p$) and an orthogonal map ($o$).
$p$ is a pure scaling map that scales each of the basis vectors by varying amounts. This gives a diagonal matrix.
$o$ is a re-orienting map that rotates/reflects the basis vectors but does not scale them (and keeps the angles between them). This is expressed as an orthogonal matrix with determinant $ pm 1$ and $ o^{*} = o^{-1} $.
So what does $ f^{*} $ do?
$ f^{*} = (op)^{*} = p^{*}o^{*} = p^{*}o^{-1}$
Which can be interpreted as: $f^{*}$ un-does the re-orientation of $ f$ and then scales the basis vectors by the same factors as $f$ does.
This is really the key to why determinants are the same.
$endgroup$
$begingroup$
I'm not sure if this makes sense. You're essentially saying that the determinant of an identity matrix is 1.
$endgroup$
– Tunococ
Aug 20 '16 at 11:36
$begingroup$
I added a different way of looking at it, hope it clarifies what I am trying to say. Note that in the first treatment, I kept the same measure of length in both vector spaces so the determinants will not all be 1.
$endgroup$
– RSS
Aug 24 '16 at 23:01
$begingroup$
My question doesn't ask for a proof. It asks for a rudimentary explanation based on the "volume spanned by the columns" point of view. I cannot really relate the matrix for $f^*f$ to column vectors geometrically, not to mention its eigenvectors. In fact, matrix multiplication is already almost too hard. I would be content, however, to hear that there's a reason why we have to resort to more difficult concepts to justify this simple-looking formula.
$endgroup$
– Tunococ
Aug 30 '16 at 11:39
$begingroup$
The algebraic proof was for this fact: both the matrix and its transpose scale the basis vectors by the same amount. Once you see that, equality of determinants is easy to see. I have also added another interpretation, probably not satisfactory but just putting it out there for completeness.
$endgroup$
– RSS
Aug 30 '16 at 13:05
$begingroup$
I'm not sure if you understand the question, but I really appreciate your attempt.
$endgroup$
– Tunococ
Aug 30 '16 at 22:49
add a comment |
$begingroup$
This is not fully geometric, but it's based on geometric considerations and I found it illuminating. It's based on the proof found in https://mathcs.clarku.edu/~ma130/determinants3.pdf. This proof is very similar to user1551's above, only presented prehaps at a more elementary level.
We want to prove that for any matrix $A$ of size $ntimes n$, $det A = det A^T$. We divide the proof to two cases.
Case 1: $rank A=r<n$. In this case the Null-space of the matrix has a positive dimension, and the co-kernel (i.e. $Nul A^T$) has the same, strictly-positive, dimension. It follows that the transformation $A$ (as well as $A^T$) "flattens" n-dimensional objects into r-dimensional objects, and thus their volume drops to zero; hence $det A=det A^T=0$.
Case 2: $rank A=r=n$. In this case the transformation $A$ is reversible (it's an isomorphism between the domain and the target), and is therefor row-equivalent to the identity matrix. In other words, we can think of $A$ as being derived from $I$ by a series of expansions (row operations of multiplication by a scalar), shears (row operatios of row addition), and reflections about the line between two axes (row operations of row exchange). Each of these can be represented with an elementary matrix so that $A=E_1 E_2 ... E_k$.
The key algebraic insight needed to complete the proof is that $(AB)^T=B^T A^T$. This makes geometric sense as the transpose transformation is working in the opposite direction to the non-transposed one, but I don't know a geometric proof of this. Given this theorem, however, it is clear that one can write $A^T = E_k^T E_{k-1}^T ... E_1^T$.
The final step is to note that for any elementary operation $det E = det E^T$. This can be understood directly from the geometric interpretation of these operations, noted above. The geometric interpretation of the determinant also naturally leads to the theorem $det AB = det A det B$. Combining these, we have
$det A = (det E_1)(det E_2)...(det E_k)=(det E_k^T)...(det E_1^T)=det A^T$
So this proof isn't completely geometric, but it's very close, and it shows very clearly the geometric meaning of the theorem: the theorem holds because we can think of $A$ as a series of elementary transformations whose transpose has the same determinant, and thus the same series transposed generates $A^T$ with the same combined determinant.
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f1319844%2fgeometric-interpretation-of-determinant-of-transpose%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
4 Answers
4
active
oldest
votes
4 Answers
4
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Write $A$ as a product of the form $E_1E_2cdots E_kDF_1F_2cdots F_ell$, where each $E_i$ (resp. $F_j$) is an elementary matrix corresponding to row (resp. column) switching or row (resp. column) addition, and $D$ is a diagonal matrix. You are done if you can explain why
- $det A=det(E_1)cdotsdet(E_k)det(D)det(F_1)cdotsdet(F_ell)$ and
- the determinant of each elementary matrix is equal to the determinant of its transpose.
But these are essentially what motivate the definition of determinant: we want to define the determinant as the volume of the parallelepiped generated by the columns of a matrix. Shears do not change the volume (thus the determinant of an elementary matrix for row/column addition is $1$) and scaling in one direction changes the volume in the same proportion (thus $det D$ is the product of $D$'s diagonal entries). Since we want to define the volume as zero in the degenerate case, it follows that together with the previous two requirements, the volume must be signed and by flipping two columns, the sign flips too (thus the determinant of an elementary matrix for row/column switching is $-1$).
$endgroup$
$begingroup$
The proof by using a sequence of shear transformations is precisely what I mentioned in my question. However, your answer did motivate me to think deeper about this proof: How would you "geometrically" explains that the determinant of each elementary matrix is equal to the determinant of its transpose? In other words, if you view columns of the matrix as vectors, why do row operations (changes of basis) affect the volume measurement the way they do?
$endgroup$
– Tunococ
Sep 23 '15 at 23:16
$begingroup$
@Tunococ Isn't that because the transpose of a shear matrix is also a shear matrix? But just how "geometric" is this is debatable. Anyway, I think the crux of the problem is that we on one hand want to define $A^T$ geometrically as the adjoint of $A$ defined on a dual space, but on the other hand we want to compare the parallelepipeds generated by the columns of $A$ and $A^T$ on the same space. Without doing some algebra, I don't think it's easy to see that the transpose of a shear is a shear.
$endgroup$
– user1551
Sep 24 '15 at 2:16
add a comment |
$begingroup$
Write $A$ as a product of the form $E_1E_2cdots E_kDF_1F_2cdots F_ell$, where each $E_i$ (resp. $F_j$) is an elementary matrix corresponding to row (resp. column) switching or row (resp. column) addition, and $D$ is a diagonal matrix. You are done if you can explain why
- $det A=det(E_1)cdotsdet(E_k)det(D)det(F_1)cdotsdet(F_ell)$ and
- the determinant of each elementary matrix is equal to the determinant of its transpose.
But these are essentially what motivate the definition of determinant: we want to define the determinant as the volume of the parallelepiped generated by the columns of a matrix. Shears do not change the volume (thus the determinant of an elementary matrix for row/column addition is $1$) and scaling in one direction changes the volume in the same proportion (thus $det D$ is the product of $D$'s diagonal entries). Since we want to define the volume as zero in the degenerate case, it follows that together with the previous two requirements, the volume must be signed and by flipping two columns, the sign flips too (thus the determinant of an elementary matrix for row/column switching is $-1$).
$endgroup$
$begingroup$
The proof by using a sequence of shear transformations is precisely what I mentioned in my question. However, your answer did motivate me to think deeper about this proof: How would you "geometrically" explains that the determinant of each elementary matrix is equal to the determinant of its transpose? In other words, if you view columns of the matrix as vectors, why do row operations (changes of basis) affect the volume measurement the way they do?
$endgroup$
– Tunococ
Sep 23 '15 at 23:16
$begingroup$
@Tunococ Isn't that because the transpose of a shear matrix is also a shear matrix? But just how "geometric" is this is debatable. Anyway, I think the crux of the problem is that we on one hand want to define $A^T$ geometrically as the adjoint of $A$ defined on a dual space, but on the other hand we want to compare the parallelepipeds generated by the columns of $A$ and $A^T$ on the same space. Without doing some algebra, I don't think it's easy to see that the transpose of a shear is a shear.
$endgroup$
– user1551
Sep 24 '15 at 2:16
add a comment |
$begingroup$
Write $A$ as a product of the form $E_1E_2cdots E_kDF_1F_2cdots F_ell$, where each $E_i$ (resp. $F_j$) is an elementary matrix corresponding to row (resp. column) switching or row (resp. column) addition, and $D$ is a diagonal matrix. You are done if you can explain why
- $det A=det(E_1)cdotsdet(E_k)det(D)det(F_1)cdotsdet(F_ell)$ and
- the determinant of each elementary matrix is equal to the determinant of its transpose.
But these are essentially what motivate the definition of determinant: we want to define the determinant as the volume of the parallelepiped generated by the columns of a matrix. Shears do not change the volume (thus the determinant of an elementary matrix for row/column addition is $1$) and scaling in one direction changes the volume in the same proportion (thus $det D$ is the product of $D$'s diagonal entries). Since we want to define the volume as zero in the degenerate case, it follows that together with the previous two requirements, the volume must be signed and by flipping two columns, the sign flips too (thus the determinant of an elementary matrix for row/column switching is $-1$).
$endgroup$
Write $A$ as a product of the form $E_1E_2cdots E_kDF_1F_2cdots F_ell$, where each $E_i$ (resp. $F_j$) is an elementary matrix corresponding to row (resp. column) switching or row (resp. column) addition, and $D$ is a diagonal matrix. You are done if you can explain why
- $det A=det(E_1)cdotsdet(E_k)det(D)det(F_1)cdotsdet(F_ell)$ and
- the determinant of each elementary matrix is equal to the determinant of its transpose.
But these are essentially what motivate the definition of determinant: we want to define the determinant as the volume of the parallelepiped generated by the columns of a matrix. Shears do not change the volume (thus the determinant of an elementary matrix for row/column addition is $1$) and scaling in one direction changes the volume in the same proportion (thus $det D$ is the product of $D$'s diagonal entries). Since we want to define the volume as zero in the degenerate case, it follows that together with the previous two requirements, the volume must be signed and by flipping two columns, the sign flips too (thus the determinant of an elementary matrix for row/column switching is $-1$).
answered Sep 23 '15 at 5:45
user1551user1551
73.1k566128
73.1k566128
$begingroup$
The proof by using a sequence of shear transformations is precisely what I mentioned in my question. However, your answer did motivate me to think deeper about this proof: How would you "geometrically" explains that the determinant of each elementary matrix is equal to the determinant of its transpose? In other words, if you view columns of the matrix as vectors, why do row operations (changes of basis) affect the volume measurement the way they do?
$endgroup$
– Tunococ
Sep 23 '15 at 23:16
$begingroup$
@Tunococ Isn't that because the transpose of a shear matrix is also a shear matrix? But just how "geometric" is this is debatable. Anyway, I think the crux of the problem is that we on one hand want to define $A^T$ geometrically as the adjoint of $A$ defined on a dual space, but on the other hand we want to compare the parallelepipeds generated by the columns of $A$ and $A^T$ on the same space. Without doing some algebra, I don't think it's easy to see that the transpose of a shear is a shear.
$endgroup$
– user1551
Sep 24 '15 at 2:16
add a comment |
$begingroup$
The proof by using a sequence of shear transformations is precisely what I mentioned in my question. However, your answer did motivate me to think deeper about this proof: How would you "geometrically" explains that the determinant of each elementary matrix is equal to the determinant of its transpose? In other words, if you view columns of the matrix as vectors, why do row operations (changes of basis) affect the volume measurement the way they do?
$endgroup$
– Tunococ
Sep 23 '15 at 23:16
$begingroup$
@Tunococ Isn't that because the transpose of a shear matrix is also a shear matrix? But just how "geometric" is this is debatable. Anyway, I think the crux of the problem is that we on one hand want to define $A^T$ geometrically as the adjoint of $A$ defined on a dual space, but on the other hand we want to compare the parallelepipeds generated by the columns of $A$ and $A^T$ on the same space. Without doing some algebra, I don't think it's easy to see that the transpose of a shear is a shear.
$endgroup$
– user1551
Sep 24 '15 at 2:16
$begingroup$
The proof by using a sequence of shear transformations is precisely what I mentioned in my question. However, your answer did motivate me to think deeper about this proof: How would you "geometrically" explains that the determinant of each elementary matrix is equal to the determinant of its transpose? In other words, if you view columns of the matrix as vectors, why do row operations (changes of basis) affect the volume measurement the way they do?
$endgroup$
– Tunococ
Sep 23 '15 at 23:16
$begingroup$
The proof by using a sequence of shear transformations is precisely what I mentioned in my question. However, your answer did motivate me to think deeper about this proof: How would you "geometrically" explains that the determinant of each elementary matrix is equal to the determinant of its transpose? In other words, if you view columns of the matrix as vectors, why do row operations (changes of basis) affect the volume measurement the way they do?
$endgroup$
– Tunococ
Sep 23 '15 at 23:16
$begingroup$
@Tunococ Isn't that because the transpose of a shear matrix is also a shear matrix? But just how "geometric" is this is debatable. Anyway, I think the crux of the problem is that we on one hand want to define $A^T$ geometrically as the adjoint of $A$ defined on a dual space, but on the other hand we want to compare the parallelepipeds generated by the columns of $A$ and $A^T$ on the same space. Without doing some algebra, I don't think it's easy to see that the transpose of a shear is a shear.
$endgroup$
– user1551
Sep 24 '15 at 2:16
$begingroup$
@Tunococ Isn't that because the transpose of a shear matrix is also a shear matrix? But just how "geometric" is this is debatable. Anyway, I think the crux of the problem is that we on one hand want to define $A^T$ geometrically as the adjoint of $A$ defined on a dual space, but on the other hand we want to compare the parallelepipeds generated by the columns of $A$ and $A^T$ on the same space. Without doing some algebra, I don't think it's easy to see that the transpose of a shear is a shear.
$endgroup$
– user1551
Sep 24 '15 at 2:16
add a comment |
$begingroup$
To prove that the two (generalized) volumes are equal is going to require some algebra -- I don't see a way around that (because the transpose is defined algebraically). This proof doesn't require the use of matrices or characteristic equations or anything, though. I just use a geometric definition of the determinant and then an algebraic formula relating a linear transformation to its adjoint (transpose).
Consider this as the geometric definition of the determinant. If $I$ is the unit $n$-cube in $Bbb R^n$, then we can define the determinant of a linear transformation $T$ as $$underline T(I)=det(T)I$$ This formula tells us that the transformation $underline T$ just scales the unit $n$-cube by its determinant.
For vectors, you might recognize that the adjoint (transpose), $overline T$, of a linear transformation, $T$, is given by $$T(v)cdot w = vcdot overline T(w)$$ However, $I$ is not really a vector, it's an object called an $n$-blade or $n$-vector. So we have to use the modified formula $$langleunderline T(A)Brangle = langle Aoverline T(B)rangle$$ where $langle Mrangle$ returns the scalar part of the multivector $M$.
So for us, it becomes $langleunderline T(I)Irangle = langle Ioverline T(I)rangle$. On the LHS, we get $$langleunderline T(I)Irangle = langle det(T)I^2rangle = det(T)I^2$$ And on the RHS, we get $$langle Ioverline T(I)rangle = langle det(overline T)I^2rangle = det(overline T)I^2$$ Therefore $$det(T) = det(overline T) square$$
$endgroup$
$begingroup$
Yes, I understand that the actual proof has to be algebraic. I'm asking for a more informal explanation that may be able to convince young students that this equality holds without going through too much algebra. By the way, your answer creates another interesting question: What is the geometric intuition of the adjoint? Maybe we can get a better intuition from the definition of adjoint than from the definition of transpose (as in flipping a matrix).
$endgroup$
– Tunococ
Sep 23 '15 at 0:19
add a comment |
$begingroup$
To prove that the two (generalized) volumes are equal is going to require some algebra -- I don't see a way around that (because the transpose is defined algebraically). This proof doesn't require the use of matrices or characteristic equations or anything, though. I just use a geometric definition of the determinant and then an algebraic formula relating a linear transformation to its adjoint (transpose).
Consider this as the geometric definition of the determinant. If $I$ is the unit $n$-cube in $Bbb R^n$, then we can define the determinant of a linear transformation $T$ as $$underline T(I)=det(T)I$$ This formula tells us that the transformation $underline T$ just scales the unit $n$-cube by its determinant.
For vectors, you might recognize that the adjoint (transpose), $overline T$, of a linear transformation, $T$, is given by $$T(v)cdot w = vcdot overline T(w)$$ However, $I$ is not really a vector, it's an object called an $n$-blade or $n$-vector. So we have to use the modified formula $$langleunderline T(A)Brangle = langle Aoverline T(B)rangle$$ where $langle Mrangle$ returns the scalar part of the multivector $M$.
So for us, it becomes $langleunderline T(I)Irangle = langle Ioverline T(I)rangle$. On the LHS, we get $$langleunderline T(I)Irangle = langle det(T)I^2rangle = det(T)I^2$$ And on the RHS, we get $$langle Ioverline T(I)rangle = langle det(overline T)I^2rangle = det(overline T)I^2$$ Therefore $$det(T) = det(overline T) square$$
$endgroup$
$begingroup$
Yes, I understand that the actual proof has to be algebraic. I'm asking for a more informal explanation that may be able to convince young students that this equality holds without going through too much algebra. By the way, your answer creates another interesting question: What is the geometric intuition of the adjoint? Maybe we can get a better intuition from the definition of adjoint than from the definition of transpose (as in flipping a matrix).
$endgroup$
– Tunococ
Sep 23 '15 at 0:19
add a comment |
$begingroup$
To prove that the two (generalized) volumes are equal is going to require some algebra -- I don't see a way around that (because the transpose is defined algebraically). This proof doesn't require the use of matrices or characteristic equations or anything, though. I just use a geometric definition of the determinant and then an algebraic formula relating a linear transformation to its adjoint (transpose).
Consider this as the geometric definition of the determinant. If $I$ is the unit $n$-cube in $Bbb R^n$, then we can define the determinant of a linear transformation $T$ as $$underline T(I)=det(T)I$$ This formula tells us that the transformation $underline T$ just scales the unit $n$-cube by its determinant.
For vectors, you might recognize that the adjoint (transpose), $overline T$, of a linear transformation, $T$, is given by $$T(v)cdot w = vcdot overline T(w)$$ However, $I$ is not really a vector, it's an object called an $n$-blade or $n$-vector. So we have to use the modified formula $$langleunderline T(A)Brangle = langle Aoverline T(B)rangle$$ where $langle Mrangle$ returns the scalar part of the multivector $M$.
So for us, it becomes $langleunderline T(I)Irangle = langle Ioverline T(I)rangle$. On the LHS, we get $$langleunderline T(I)Irangle = langle det(T)I^2rangle = det(T)I^2$$ And on the RHS, we get $$langle Ioverline T(I)rangle = langle det(overline T)I^2rangle = det(overline T)I^2$$ Therefore $$det(T) = det(overline T) square$$
$endgroup$
To prove that the two (generalized) volumes are equal is going to require some algebra -- I don't see a way around that (because the transpose is defined algebraically). This proof doesn't require the use of matrices or characteristic equations or anything, though. I just use a geometric definition of the determinant and then an algebraic formula relating a linear transformation to its adjoint (transpose).
Consider this as the geometric definition of the determinant. If $I$ is the unit $n$-cube in $Bbb R^n$, then we can define the determinant of a linear transformation $T$ as $$underline T(I)=det(T)I$$ This formula tells us that the transformation $underline T$ just scales the unit $n$-cube by its determinant.
For vectors, you might recognize that the adjoint (transpose), $overline T$, of a linear transformation, $T$, is given by $$T(v)cdot w = vcdot overline T(w)$$ However, $I$ is not really a vector, it's an object called an $n$-blade or $n$-vector. So we have to use the modified formula $$langleunderline T(A)Brangle = langle Aoverline T(B)rangle$$ where $langle Mrangle$ returns the scalar part of the multivector $M$.
So for us, it becomes $langleunderline T(I)Irangle = langle Ioverline T(I)rangle$. On the LHS, we get $$langleunderline T(I)Irangle = langle det(T)I^2rangle = det(T)I^2$$ And on the RHS, we get $$langle Ioverline T(I)rangle = langle det(overline T)I^2rangle = det(overline T)I^2$$ Therefore $$det(T) = det(overline T) square$$
edited Sep 22 '15 at 23:47
answered Sep 22 '15 at 21:49
got it--thanksgot it--thanks
2,079920
2,079920
$begingroup$
Yes, I understand that the actual proof has to be algebraic. I'm asking for a more informal explanation that may be able to convince young students that this equality holds without going through too much algebra. By the way, your answer creates another interesting question: What is the geometric intuition of the adjoint? Maybe we can get a better intuition from the definition of adjoint than from the definition of transpose (as in flipping a matrix).
$endgroup$
– Tunococ
Sep 23 '15 at 0:19
add a comment |
$begingroup$
Yes, I understand that the actual proof has to be algebraic. I'm asking for a more informal explanation that may be able to convince young students that this equality holds without going through too much algebra. By the way, your answer creates another interesting question: What is the geometric intuition of the adjoint? Maybe we can get a better intuition from the definition of adjoint than from the definition of transpose (as in flipping a matrix).
$endgroup$
– Tunococ
Sep 23 '15 at 0:19
$begingroup$
Yes, I understand that the actual proof has to be algebraic. I'm asking for a more informal explanation that may be able to convince young students that this equality holds without going through too much algebra. By the way, your answer creates another interesting question: What is the geometric intuition of the adjoint? Maybe we can get a better intuition from the definition of adjoint than from the definition of transpose (as in flipping a matrix).
$endgroup$
– Tunococ
Sep 23 '15 at 0:19
$begingroup$
Yes, I understand that the actual proof has to be algebraic. I'm asking for a more informal explanation that may be able to convince young students that this equality holds without going through too much algebra. By the way, your answer creates another interesting question: What is the geometric intuition of the adjoint? Maybe we can get a better intuition from the definition of adjoint than from the definition of transpose (as in flipping a matrix).
$endgroup$
– Tunococ
Sep 23 '15 at 0:19
add a comment |
$begingroup$
I think there is a geometrical interpretation but it requires a bit of imagination w.r.t inner product spaces. Some notation to start:
$ f: U rightarrow V $ is a linear map and $ f^{*} : V rightarrow U $ is it's adjoint.
Let $ e_1 ..e_n s $ be an orthonormal basis for $U$. Considering only the non-degenerate case, we can see that $ frac{f(e_i)}{lVert f(e_i) rVert} = e^{'}_i $ form a basis for $V$.
Now here is the trick: define $ e^{'}_1..e^{'}_n $ to be an orthonormal basis for $V$. From the perspective of $U$, $e^{'}_is$ may not be orthonormal, but when computing inside of $V$, we can define it to be.
Physically speaking, the space $V$ appears deformed, but only as seen from $U$ (and vice-versa).
From the definition of the adjoint, for any $u in U$ and $ v in V$, we have
$ f(u)cdot v = f^{*}(v)cdot u $
Note that the L.H.S is being computed in $V$, while the R.H.S is being computed in $U$. This equation also holds for $e_is $ and $e^{'}_js$.
$ f(e_i)cdot e^{'}_j = f^{*}(e^{'}_j)cdot e_i$
But the L.H.S is zero for any $ i neq j $!
$ f^{*}(e^{'}_i)cdot e_i = f(e_i)cdot e^{'}_i = lVert f(e_i) rVert $
$ f^{*}(e^{'}_i)cdot e_j = 0$ $ (i neq j) $
Now we come to the crux of the argument. For this special choice of bases, $ f^{*} $ dilates the orthonormal basis vectors $e^{'}_i$ of $V$ by the same amount as $ f$ dilates the orthonormal basis vectors $ e_i$ of $U$. By keeping the order same, it also preserves the orientation.
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
For this reason, $f^{*}$ maps the unit 'volume' of $V$ to the same volume as $f$ does the unit 'volume' of $U$.
Edit 1:
The above treatment is admittedly not very rigorous, to start with determinants are usually defined for linear operators $ f : V rightarrow V $ whereas I stretched the definition to $ f : U rightarrow V $. But here is an alternate way to see the same thing:
for any $ f : V rightarrow V $, the composite transform $ f^{*}f : V rightarrow V $ is a symmetric linear transform (or Matrix), which has an orthornormal eigenvector basis $ e_1, e_2, .. e_n $ with real eigenvalues $ lambda_1, lambda_1,.. lambda_n$
Now consider the $ f(e_i)'s $:
$ f(e_i)cdot f(e_j) = f^{*}f(e_i)cdot e_j = lambda_i e_icdot e_j $
Therefore, $ f(e_i)s $ also form an orthornormal basis like before (except this time not by fiat). Also:
$ f(e_i).f(e_i) = lVert f(e_i) rVert^2 = lambda_i $
Which is to say $f$ scales $ e_i$ by $ sqrt{lambda_i}$. Now set the basis $ e^{'}_i = frac{f(e_i)}{sqrt{lambda_i}} $ for $ f^{*} $ and check the scaling:
$ f^{*}(e^{'}_i) = frac{f^{*}f(e_i)}{sqrt{lambda_i}} = sqrt{lambda_i}e_i$
So again,
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
Edit 2:
Another way to look at this is through the geometrical interpretation of the adjoint itself.
Every linear map, when seen in a certain special basis (as shown above), can be thought of as a composition of two maps: A positive semidefinite map ($p$) and an orthogonal map ($o$).
$p$ is a pure scaling map that scales each of the basis vectors by varying amounts. This gives a diagonal matrix.
$o$ is a re-orienting map that rotates/reflects the basis vectors but does not scale them (and keeps the angles between them). This is expressed as an orthogonal matrix with determinant $ pm 1$ and $ o^{*} = o^{-1} $.
So what does $ f^{*} $ do?
$ f^{*} = (op)^{*} = p^{*}o^{*} = p^{*}o^{-1}$
Which can be interpreted as: $f^{*}$ un-does the re-orientation of $ f$ and then scales the basis vectors by the same factors as $f$ does.
This is really the key to why determinants are the same.
$endgroup$
$begingroup$
I'm not sure if this makes sense. You're essentially saying that the determinant of an identity matrix is 1.
$endgroup$
– Tunococ
Aug 20 '16 at 11:36
$begingroup$
I added a different way of looking at it, hope it clarifies what I am trying to say. Note that in the first treatment, I kept the same measure of length in both vector spaces so the determinants will not all be 1.
$endgroup$
– RSS
Aug 24 '16 at 23:01
$begingroup$
My question doesn't ask for a proof. It asks for a rudimentary explanation based on the "volume spanned by the columns" point of view. I cannot really relate the matrix for $f^*f$ to column vectors geometrically, not to mention its eigenvectors. In fact, matrix multiplication is already almost too hard. I would be content, however, to hear that there's a reason why we have to resort to more difficult concepts to justify this simple-looking formula.
$endgroup$
– Tunococ
Aug 30 '16 at 11:39
$begingroup$
The algebraic proof was for this fact: both the matrix and its transpose scale the basis vectors by the same amount. Once you see that, equality of determinants is easy to see. I have also added another interpretation, probably not satisfactory but just putting it out there for completeness.
$endgroup$
– RSS
Aug 30 '16 at 13:05
$begingroup$
I'm not sure if you understand the question, but I really appreciate your attempt.
$endgroup$
– Tunococ
Aug 30 '16 at 22:49
add a comment |
$begingroup$
I think there is a geometrical interpretation but it requires a bit of imagination w.r.t inner product spaces. Some notation to start:
$ f: U rightarrow V $ is a linear map and $ f^{*} : V rightarrow U $ is it's adjoint.
Let $ e_1 ..e_n s $ be an orthonormal basis for $U$. Considering only the non-degenerate case, we can see that $ frac{f(e_i)}{lVert f(e_i) rVert} = e^{'}_i $ form a basis for $V$.
Now here is the trick: define $ e^{'}_1..e^{'}_n $ to be an orthonormal basis for $V$. From the perspective of $U$, $e^{'}_is$ may not be orthonormal, but when computing inside of $V$, we can define it to be.
Physically speaking, the space $V$ appears deformed, but only as seen from $U$ (and vice-versa).
From the definition of the adjoint, for any $u in U$ and $ v in V$, we have
$ f(u)cdot v = f^{*}(v)cdot u $
Note that the L.H.S is being computed in $V$, while the R.H.S is being computed in $U$. This equation also holds for $e_is $ and $e^{'}_js$.
$ f(e_i)cdot e^{'}_j = f^{*}(e^{'}_j)cdot e_i$
But the L.H.S is zero for any $ i neq j $!
$ f^{*}(e^{'}_i)cdot e_i = f(e_i)cdot e^{'}_i = lVert f(e_i) rVert $
$ f^{*}(e^{'}_i)cdot e_j = 0$ $ (i neq j) $
Now we come to the crux of the argument. For this special choice of bases, $ f^{*} $ dilates the orthonormal basis vectors $e^{'}_i$ of $V$ by the same amount as $ f$ dilates the orthonormal basis vectors $ e_i$ of $U$. By keeping the order same, it also preserves the orientation.
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
For this reason, $f^{*}$ maps the unit 'volume' of $V$ to the same volume as $f$ does the unit 'volume' of $U$.
Edit 1:
The above treatment is admittedly not very rigorous, to start with determinants are usually defined for linear operators $ f : V rightarrow V $ whereas I stretched the definition to $ f : U rightarrow V $. But here is an alternate way to see the same thing:
for any $ f : V rightarrow V $, the composite transform $ f^{*}f : V rightarrow V $ is a symmetric linear transform (or Matrix), which has an orthornormal eigenvector basis $ e_1, e_2, .. e_n $ with real eigenvalues $ lambda_1, lambda_1,.. lambda_n$
Now consider the $ f(e_i)'s $:
$ f(e_i)cdot f(e_j) = f^{*}f(e_i)cdot e_j = lambda_i e_icdot e_j $
Therefore, $ f(e_i)s $ also form an orthornormal basis like before (except this time not by fiat). Also:
$ f(e_i).f(e_i) = lVert f(e_i) rVert^2 = lambda_i $
Which is to say $f$ scales $ e_i$ by $ sqrt{lambda_i}$. Now set the basis $ e^{'}_i = frac{f(e_i)}{sqrt{lambda_i}} $ for $ f^{*} $ and check the scaling:
$ f^{*}(e^{'}_i) = frac{f^{*}f(e_i)}{sqrt{lambda_i}} = sqrt{lambda_i}e_i$
So again,
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
Edit 2:
Another way to look at this is through the geometrical interpretation of the adjoint itself.
Every linear map, when seen in a certain special basis (as shown above), can be thought of as a composition of two maps: A positive semidefinite map ($p$) and an orthogonal map ($o$).
$p$ is a pure scaling map that scales each of the basis vectors by varying amounts. This gives a diagonal matrix.
$o$ is a re-orienting map that rotates/reflects the basis vectors but does not scale them (and keeps the angles between them). This is expressed as an orthogonal matrix with determinant $ pm 1$ and $ o^{*} = o^{-1} $.
So what does $ f^{*} $ do?
$ f^{*} = (op)^{*} = p^{*}o^{*} = p^{*}o^{-1}$
Which can be interpreted as: $f^{*}$ un-does the re-orientation of $ f$ and then scales the basis vectors by the same factors as $f$ does.
This is really the key to why determinants are the same.
$endgroup$
$begingroup$
I'm not sure if this makes sense. You're essentially saying that the determinant of an identity matrix is 1.
$endgroup$
– Tunococ
Aug 20 '16 at 11:36
$begingroup$
I added a different way of looking at it, hope it clarifies what I am trying to say. Note that in the first treatment, I kept the same measure of length in both vector spaces so the determinants will not all be 1.
$endgroup$
– RSS
Aug 24 '16 at 23:01
$begingroup$
My question doesn't ask for a proof. It asks for a rudimentary explanation based on the "volume spanned by the columns" point of view. I cannot really relate the matrix for $f^*f$ to column vectors geometrically, not to mention its eigenvectors. In fact, matrix multiplication is already almost too hard. I would be content, however, to hear that there's a reason why we have to resort to more difficult concepts to justify this simple-looking formula.
$endgroup$
– Tunococ
Aug 30 '16 at 11:39
$begingroup$
The algebraic proof was for this fact: both the matrix and its transpose scale the basis vectors by the same amount. Once you see that, equality of determinants is easy to see. I have also added another interpretation, probably not satisfactory but just putting it out there for completeness.
$endgroup$
– RSS
Aug 30 '16 at 13:05
$begingroup$
I'm not sure if you understand the question, but I really appreciate your attempt.
$endgroup$
– Tunococ
Aug 30 '16 at 22:49
add a comment |
$begingroup$
I think there is a geometrical interpretation but it requires a bit of imagination w.r.t inner product spaces. Some notation to start:
$ f: U rightarrow V $ is a linear map and $ f^{*} : V rightarrow U $ is it's adjoint.
Let $ e_1 ..e_n s $ be an orthonormal basis for $U$. Considering only the non-degenerate case, we can see that $ frac{f(e_i)}{lVert f(e_i) rVert} = e^{'}_i $ form a basis for $V$.
Now here is the trick: define $ e^{'}_1..e^{'}_n $ to be an orthonormal basis for $V$. From the perspective of $U$, $e^{'}_is$ may not be orthonormal, but when computing inside of $V$, we can define it to be.
Physically speaking, the space $V$ appears deformed, but only as seen from $U$ (and vice-versa).
From the definition of the adjoint, for any $u in U$ and $ v in V$, we have
$ f(u)cdot v = f^{*}(v)cdot u $
Note that the L.H.S is being computed in $V$, while the R.H.S is being computed in $U$. This equation also holds for $e_is $ and $e^{'}_js$.
$ f(e_i)cdot e^{'}_j = f^{*}(e^{'}_j)cdot e_i$
But the L.H.S is zero for any $ i neq j $!
$ f^{*}(e^{'}_i)cdot e_i = f(e_i)cdot e^{'}_i = lVert f(e_i) rVert $
$ f^{*}(e^{'}_i)cdot e_j = 0$ $ (i neq j) $
Now we come to the crux of the argument. For this special choice of bases, $ f^{*} $ dilates the orthonormal basis vectors $e^{'}_i$ of $V$ by the same amount as $ f$ dilates the orthonormal basis vectors $ e_i$ of $U$. By keeping the order same, it also preserves the orientation.
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
For this reason, $f^{*}$ maps the unit 'volume' of $V$ to the same volume as $f$ does the unit 'volume' of $U$.
Edit 1:
The above treatment is admittedly not very rigorous, to start with determinants are usually defined for linear operators $ f : V rightarrow V $ whereas I stretched the definition to $ f : U rightarrow V $. But here is an alternate way to see the same thing:
for any $ f : V rightarrow V $, the composite transform $ f^{*}f : V rightarrow V $ is a symmetric linear transform (or Matrix), which has an orthornormal eigenvector basis $ e_1, e_2, .. e_n $ with real eigenvalues $ lambda_1, lambda_1,.. lambda_n$
Now consider the $ f(e_i)'s $:
$ f(e_i)cdot f(e_j) = f^{*}f(e_i)cdot e_j = lambda_i e_icdot e_j $
Therefore, $ f(e_i)s $ also form an orthornormal basis like before (except this time not by fiat). Also:
$ f(e_i).f(e_i) = lVert f(e_i) rVert^2 = lambda_i $
Which is to say $f$ scales $ e_i$ by $ sqrt{lambda_i}$. Now set the basis $ e^{'}_i = frac{f(e_i)}{sqrt{lambda_i}} $ for $ f^{*} $ and check the scaling:
$ f^{*}(e^{'}_i) = frac{f^{*}f(e_i)}{sqrt{lambda_i}} = sqrt{lambda_i}e_i$
So again,
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
Edit 2:
Another way to look at this is through the geometrical interpretation of the adjoint itself.
Every linear map, when seen in a certain special basis (as shown above), can be thought of as a composition of two maps: A positive semidefinite map ($p$) and an orthogonal map ($o$).
$p$ is a pure scaling map that scales each of the basis vectors by varying amounts. This gives a diagonal matrix.
$o$ is a re-orienting map that rotates/reflects the basis vectors but does not scale them (and keeps the angles between them). This is expressed as an orthogonal matrix with determinant $ pm 1$ and $ o^{*} = o^{-1} $.
So what does $ f^{*} $ do?
$ f^{*} = (op)^{*} = p^{*}o^{*} = p^{*}o^{-1}$
Which can be interpreted as: $f^{*}$ un-does the re-orientation of $ f$ and then scales the basis vectors by the same factors as $f$ does.
This is really the key to why determinants are the same.
$endgroup$
I think there is a geometrical interpretation but it requires a bit of imagination w.r.t inner product spaces. Some notation to start:
$ f: U rightarrow V $ is a linear map and $ f^{*} : V rightarrow U $ is it's adjoint.
Let $ e_1 ..e_n s $ be an orthonormal basis for $U$. Considering only the non-degenerate case, we can see that $ frac{f(e_i)}{lVert f(e_i) rVert} = e^{'}_i $ form a basis for $V$.
Now here is the trick: define $ e^{'}_1..e^{'}_n $ to be an orthonormal basis for $V$. From the perspective of $U$, $e^{'}_is$ may not be orthonormal, but when computing inside of $V$, we can define it to be.
Physically speaking, the space $V$ appears deformed, but only as seen from $U$ (and vice-versa).
From the definition of the adjoint, for any $u in U$ and $ v in V$, we have
$ f(u)cdot v = f^{*}(v)cdot u $
Note that the L.H.S is being computed in $V$, while the R.H.S is being computed in $U$. This equation also holds for $e_is $ and $e^{'}_js$.
$ f(e_i)cdot e^{'}_j = f^{*}(e^{'}_j)cdot e_i$
But the L.H.S is zero for any $ i neq j $!
$ f^{*}(e^{'}_i)cdot e_i = f(e_i)cdot e^{'}_i = lVert f(e_i) rVert $
$ f^{*}(e^{'}_i)cdot e_j = 0$ $ (i neq j) $
Now we come to the crux of the argument. For this special choice of bases, $ f^{*} $ dilates the orthonormal basis vectors $e^{'}_i$ of $V$ by the same amount as $ f$ dilates the orthonormal basis vectors $ e_i$ of $U$. By keeping the order same, it also preserves the orientation.
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
For this reason, $f^{*}$ maps the unit 'volume' of $V$ to the same volume as $f$ does the unit 'volume' of $U$.
Edit 1:
The above treatment is admittedly not very rigorous, to start with determinants are usually defined for linear operators $ f : V rightarrow V $ whereas I stretched the definition to $ f : U rightarrow V $. But here is an alternate way to see the same thing:
for any $ f : V rightarrow V $, the composite transform $ f^{*}f : V rightarrow V $ is a symmetric linear transform (or Matrix), which has an orthornormal eigenvector basis $ e_1, e_2, .. e_n $ with real eigenvalues $ lambda_1, lambda_1,.. lambda_n$
Now consider the $ f(e_i)'s $:
$ f(e_i)cdot f(e_j) = f^{*}f(e_i)cdot e_j = lambda_i e_icdot e_j $
Therefore, $ f(e_i)s $ also form an orthornormal basis like before (except this time not by fiat). Also:
$ f(e_i).f(e_i) = lVert f(e_i) rVert^2 = lambda_i $
Which is to say $f$ scales $ e_i$ by $ sqrt{lambda_i}$. Now set the basis $ e^{'}_i = frac{f(e_i)}{sqrt{lambda_i}} $ for $ f^{*} $ and check the scaling:
$ f^{*}(e^{'}_i) = frac{f^{*}f(e_i)}{sqrt{lambda_i}} = sqrt{lambda_i}e_i$
So again,
$ lVert f(e_i) rVert = lVert f^{*}(e^{'}_i) rVert $
Edit 2:
Another way to look at this is through the geometrical interpretation of the adjoint itself.
Every linear map, when seen in a certain special basis (as shown above), can be thought of as a composition of two maps: A positive semidefinite map ($p$) and an orthogonal map ($o$).
$p$ is a pure scaling map that scales each of the basis vectors by varying amounts. This gives a diagonal matrix.
$o$ is a re-orienting map that rotates/reflects the basis vectors but does not scale them (and keeps the angles between them). This is expressed as an orthogonal matrix with determinant $ pm 1$ and $ o^{*} = o^{-1} $.
So what does $ f^{*} $ do?
$ f^{*} = (op)^{*} = p^{*}o^{*} = p^{*}o^{-1}$
Which can be interpreted as: $f^{*}$ un-does the re-orientation of $ f$ and then scales the basis vectors by the same factors as $f$ does.
This is really the key to why determinants are the same.
edited Apr 13 '17 at 12:19
Community♦
1
1
answered Aug 19 '16 at 20:33
RSSRSS
32029
32029
$begingroup$
I'm not sure if this makes sense. You're essentially saying that the determinant of an identity matrix is 1.
$endgroup$
– Tunococ
Aug 20 '16 at 11:36
$begingroup$
I added a different way of looking at it, hope it clarifies what I am trying to say. Note that in the first treatment, I kept the same measure of length in both vector spaces so the determinants will not all be 1.
$endgroup$
– RSS
Aug 24 '16 at 23:01
$begingroup$
My question doesn't ask for a proof. It asks for a rudimentary explanation based on the "volume spanned by the columns" point of view. I cannot really relate the matrix for $f^*f$ to column vectors geometrically, not to mention its eigenvectors. In fact, matrix multiplication is already almost too hard. I would be content, however, to hear that there's a reason why we have to resort to more difficult concepts to justify this simple-looking formula.
$endgroup$
– Tunococ
Aug 30 '16 at 11:39
$begingroup$
The algebraic proof was for this fact: both the matrix and its transpose scale the basis vectors by the same amount. Once you see that, equality of determinants is easy to see. I have also added another interpretation, probably not satisfactory but just putting it out there for completeness.
$endgroup$
– RSS
Aug 30 '16 at 13:05
$begingroup$
I'm not sure if you understand the question, but I really appreciate your attempt.
$endgroup$
– Tunococ
Aug 30 '16 at 22:49
add a comment |
$begingroup$
I'm not sure if this makes sense. You're essentially saying that the determinant of an identity matrix is 1.
$endgroup$
– Tunococ
Aug 20 '16 at 11:36
$begingroup$
I added a different way of looking at it, hope it clarifies what I am trying to say. Note that in the first treatment, I kept the same measure of length in both vector spaces so the determinants will not all be 1.
$endgroup$
– RSS
Aug 24 '16 at 23:01
$begingroup$
My question doesn't ask for a proof. It asks for a rudimentary explanation based on the "volume spanned by the columns" point of view. I cannot really relate the matrix for $f^*f$ to column vectors geometrically, not to mention its eigenvectors. In fact, matrix multiplication is already almost too hard. I would be content, however, to hear that there's a reason why we have to resort to more difficult concepts to justify this simple-looking formula.
$endgroup$
– Tunococ
Aug 30 '16 at 11:39
$begingroup$
The algebraic proof was for this fact: both the matrix and its transpose scale the basis vectors by the same amount. Once you see that, equality of determinants is easy to see. I have also added another interpretation, probably not satisfactory but just putting it out there for completeness.
$endgroup$
– RSS
Aug 30 '16 at 13:05
$begingroup$
I'm not sure if you understand the question, but I really appreciate your attempt.
$endgroup$
– Tunococ
Aug 30 '16 at 22:49
$begingroup$
I'm not sure if this makes sense. You're essentially saying that the determinant of an identity matrix is 1.
$endgroup$
– Tunococ
Aug 20 '16 at 11:36
$begingroup$
I'm not sure if this makes sense. You're essentially saying that the determinant of an identity matrix is 1.
$endgroup$
– Tunococ
Aug 20 '16 at 11:36
$begingroup$
I added a different way of looking at it, hope it clarifies what I am trying to say. Note that in the first treatment, I kept the same measure of length in both vector spaces so the determinants will not all be 1.
$endgroup$
– RSS
Aug 24 '16 at 23:01
$begingroup$
I added a different way of looking at it, hope it clarifies what I am trying to say. Note that in the first treatment, I kept the same measure of length in both vector spaces so the determinants will not all be 1.
$endgroup$
– RSS
Aug 24 '16 at 23:01
$begingroup$
My question doesn't ask for a proof. It asks for a rudimentary explanation based on the "volume spanned by the columns" point of view. I cannot really relate the matrix for $f^*f$ to column vectors geometrically, not to mention its eigenvectors. In fact, matrix multiplication is already almost too hard. I would be content, however, to hear that there's a reason why we have to resort to more difficult concepts to justify this simple-looking formula.
$endgroup$
– Tunococ
Aug 30 '16 at 11:39
$begingroup$
My question doesn't ask for a proof. It asks for a rudimentary explanation based on the "volume spanned by the columns" point of view. I cannot really relate the matrix for $f^*f$ to column vectors geometrically, not to mention its eigenvectors. In fact, matrix multiplication is already almost too hard. I would be content, however, to hear that there's a reason why we have to resort to more difficult concepts to justify this simple-looking formula.
$endgroup$
– Tunococ
Aug 30 '16 at 11:39
$begingroup$
The algebraic proof was for this fact: both the matrix and its transpose scale the basis vectors by the same amount. Once you see that, equality of determinants is easy to see. I have also added another interpretation, probably not satisfactory but just putting it out there for completeness.
$endgroup$
– RSS
Aug 30 '16 at 13:05
$begingroup$
The algebraic proof was for this fact: both the matrix and its transpose scale the basis vectors by the same amount. Once you see that, equality of determinants is easy to see. I have also added another interpretation, probably not satisfactory but just putting it out there for completeness.
$endgroup$
– RSS
Aug 30 '16 at 13:05
$begingroup$
I'm not sure if you understand the question, but I really appreciate your attempt.
$endgroup$
– Tunococ
Aug 30 '16 at 22:49
$begingroup$
I'm not sure if you understand the question, but I really appreciate your attempt.
$endgroup$
– Tunococ
Aug 30 '16 at 22:49
add a comment |
$begingroup$
This is not fully geometric, but it's based on geometric considerations and I found it illuminating. It's based on the proof found in https://mathcs.clarku.edu/~ma130/determinants3.pdf. This proof is very similar to user1551's above, only presented prehaps at a more elementary level.
We want to prove that for any matrix $A$ of size $ntimes n$, $det A = det A^T$. We divide the proof to two cases.
Case 1: $rank A=r<n$. In this case the Null-space of the matrix has a positive dimension, and the co-kernel (i.e. $Nul A^T$) has the same, strictly-positive, dimension. It follows that the transformation $A$ (as well as $A^T$) "flattens" n-dimensional objects into r-dimensional objects, and thus their volume drops to zero; hence $det A=det A^T=0$.
Case 2: $rank A=r=n$. In this case the transformation $A$ is reversible (it's an isomorphism between the domain and the target), and is therefor row-equivalent to the identity matrix. In other words, we can think of $A$ as being derived from $I$ by a series of expansions (row operations of multiplication by a scalar), shears (row operatios of row addition), and reflections about the line between two axes (row operations of row exchange). Each of these can be represented with an elementary matrix so that $A=E_1 E_2 ... E_k$.
The key algebraic insight needed to complete the proof is that $(AB)^T=B^T A^T$. This makes geometric sense as the transpose transformation is working in the opposite direction to the non-transposed one, but I don't know a geometric proof of this. Given this theorem, however, it is clear that one can write $A^T = E_k^T E_{k-1}^T ... E_1^T$.
The final step is to note that for any elementary operation $det E = det E^T$. This can be understood directly from the geometric interpretation of these operations, noted above. The geometric interpretation of the determinant also naturally leads to the theorem $det AB = det A det B$. Combining these, we have
$det A = (det E_1)(det E_2)...(det E_k)=(det E_k^T)...(det E_1^T)=det A^T$
So this proof isn't completely geometric, but it's very close, and it shows very clearly the geometric meaning of the theorem: the theorem holds because we can think of $A$ as a series of elementary transformations whose transpose has the same determinant, and thus the same series transposed generates $A^T$ with the same combined determinant.
$endgroup$
add a comment |
$begingroup$
This is not fully geometric, but it's based on geometric considerations and I found it illuminating. It's based on the proof found in https://mathcs.clarku.edu/~ma130/determinants3.pdf. This proof is very similar to user1551's above, only presented prehaps at a more elementary level.
We want to prove that for any matrix $A$ of size $ntimes n$, $det A = det A^T$. We divide the proof to two cases.
Case 1: $rank A=r<n$. In this case the Null-space of the matrix has a positive dimension, and the co-kernel (i.e. $Nul A^T$) has the same, strictly-positive, dimension. It follows that the transformation $A$ (as well as $A^T$) "flattens" n-dimensional objects into r-dimensional objects, and thus their volume drops to zero; hence $det A=det A^T=0$.
Case 2: $rank A=r=n$. In this case the transformation $A$ is reversible (it's an isomorphism between the domain and the target), and is therefor row-equivalent to the identity matrix. In other words, we can think of $A$ as being derived from $I$ by a series of expansions (row operations of multiplication by a scalar), shears (row operatios of row addition), and reflections about the line between two axes (row operations of row exchange). Each of these can be represented with an elementary matrix so that $A=E_1 E_2 ... E_k$.
The key algebraic insight needed to complete the proof is that $(AB)^T=B^T A^T$. This makes geometric sense as the transpose transformation is working in the opposite direction to the non-transposed one, but I don't know a geometric proof of this. Given this theorem, however, it is clear that one can write $A^T = E_k^T E_{k-1}^T ... E_1^T$.
The final step is to note that for any elementary operation $det E = det E^T$. This can be understood directly from the geometric interpretation of these operations, noted above. The geometric interpretation of the determinant also naturally leads to the theorem $det AB = det A det B$. Combining these, we have
$det A = (det E_1)(det E_2)...(det E_k)=(det E_k^T)...(det E_1^T)=det A^T$
So this proof isn't completely geometric, but it's very close, and it shows very clearly the geometric meaning of the theorem: the theorem holds because we can think of $A$ as a series of elementary transformations whose transpose has the same determinant, and thus the same series transposed generates $A^T$ with the same combined determinant.
$endgroup$
add a comment |
$begingroup$
This is not fully geometric, but it's based on geometric considerations and I found it illuminating. It's based on the proof found in https://mathcs.clarku.edu/~ma130/determinants3.pdf. This proof is very similar to user1551's above, only presented prehaps at a more elementary level.
We want to prove that for any matrix $A$ of size $ntimes n$, $det A = det A^T$. We divide the proof to two cases.
Case 1: $rank A=r<n$. In this case the Null-space of the matrix has a positive dimension, and the co-kernel (i.e. $Nul A^T$) has the same, strictly-positive, dimension. It follows that the transformation $A$ (as well as $A^T$) "flattens" n-dimensional objects into r-dimensional objects, and thus their volume drops to zero; hence $det A=det A^T=0$.
Case 2: $rank A=r=n$. In this case the transformation $A$ is reversible (it's an isomorphism between the domain and the target), and is therefor row-equivalent to the identity matrix. In other words, we can think of $A$ as being derived from $I$ by a series of expansions (row operations of multiplication by a scalar), shears (row operatios of row addition), and reflections about the line between two axes (row operations of row exchange). Each of these can be represented with an elementary matrix so that $A=E_1 E_2 ... E_k$.
The key algebraic insight needed to complete the proof is that $(AB)^T=B^T A^T$. This makes geometric sense as the transpose transformation is working in the opposite direction to the non-transposed one, but I don't know a geometric proof of this. Given this theorem, however, it is clear that one can write $A^T = E_k^T E_{k-1}^T ... E_1^T$.
The final step is to note that for any elementary operation $det E = det E^T$. This can be understood directly from the geometric interpretation of these operations, noted above. The geometric interpretation of the determinant also naturally leads to the theorem $det AB = det A det B$. Combining these, we have
$det A = (det E_1)(det E_2)...(det E_k)=(det E_k^T)...(det E_1^T)=det A^T$
So this proof isn't completely geometric, but it's very close, and it shows very clearly the geometric meaning of the theorem: the theorem holds because we can think of $A$ as a series of elementary transformations whose transpose has the same determinant, and thus the same series transposed generates $A^T$ with the same combined determinant.
$endgroup$
This is not fully geometric, but it's based on geometric considerations and I found it illuminating. It's based on the proof found in https://mathcs.clarku.edu/~ma130/determinants3.pdf. This proof is very similar to user1551's above, only presented prehaps at a more elementary level.
We want to prove that for any matrix $A$ of size $ntimes n$, $det A = det A^T$. We divide the proof to two cases.
Case 1: $rank A=r<n$. In this case the Null-space of the matrix has a positive dimension, and the co-kernel (i.e. $Nul A^T$) has the same, strictly-positive, dimension. It follows that the transformation $A$ (as well as $A^T$) "flattens" n-dimensional objects into r-dimensional objects, and thus their volume drops to zero; hence $det A=det A^T=0$.
Case 2: $rank A=r=n$. In this case the transformation $A$ is reversible (it's an isomorphism between the domain and the target), and is therefor row-equivalent to the identity matrix. In other words, we can think of $A$ as being derived from $I$ by a series of expansions (row operations of multiplication by a scalar), shears (row operatios of row addition), and reflections about the line between two axes (row operations of row exchange). Each of these can be represented with an elementary matrix so that $A=E_1 E_2 ... E_k$.
The key algebraic insight needed to complete the proof is that $(AB)^T=B^T A^T$. This makes geometric sense as the transpose transformation is working in the opposite direction to the non-transposed one, but I don't know a geometric proof of this. Given this theorem, however, it is clear that one can write $A^T = E_k^T E_{k-1}^T ... E_1^T$.
The final step is to note that for any elementary operation $det E = det E^T$. This can be understood directly from the geometric interpretation of these operations, noted above. The geometric interpretation of the determinant also naturally leads to the theorem $det AB = det A det B$. Combining these, we have
$det A = (det E_1)(det E_2)...(det E_k)=(det E_k^T)...(det E_1^T)=det A^T$
So this proof isn't completely geometric, but it's very close, and it shows very clearly the geometric meaning of the theorem: the theorem holds because we can think of $A$ as a series of elementary transformations whose transpose has the same determinant, and thus the same series transposed generates $A^T$ with the same combined determinant.
edited Jan 6 '18 at 13:16
answered Jan 6 '18 at 13:08
PhysicsTeacherPhysicsTeacher
1142
1142
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f1319844%2fgeometric-interpretation-of-determinant-of-transpose%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
This looks like a duplicate of this question
$endgroup$
– preferred_anon
Jun 10 '15 at 12:17