You're reading the public-facing archive of the Category Theory Zulip server.
To join the server you need an invite. Anybody can get an invite by contacting Matteo Capucci at name dot surname at gmail dot com.
For all things related to this archive refer to the same person.
I never took a course on commutative algebra because in grad school I was mainly interested in mathematical physics, so I'm just finally now learning Nakayama's lemma - because I need it for something.
There are many alternative formulations, some perhaps easier to remember than others. MathOverflow someone gave this nice mnemonic for it which happens to match what I actually need:
What does this mean? It means:
Nakayama's Lemma. Let be an ideal in a commutative ring and let be a finitely generated -module. If then there exists such that for all .
This instantly implies something I wanted to know: if is an ideal in a commutative ring with , then for some with .
Starting from this simple formulation I'm finally getting more interested in the many alternative formulations and consequences, some of which sound more conceptual, like
If is a surjective endomorphism of a finitely generated -module, then is an isomorphism.
or
Any basis of the fiber of a coherent sheaf at a point extends to a minimal generating set of local sections.
or
A finitely generated module over a local ring is projective only if it is free.
John Baez said:
This instantly implies something I wanted to know: if is an ideal in a commutative ring with , then for some with .
I don't understand how you get this instantly. And is it even true? I've seen it written on the internet with the additional hypothesis that is finitely generated.
The ideal does have to be finitely generated to apply Nakayama's Lemma, indeed. But otherwise it's immediate, as:
John Baez said:
Starting from this simple formulation I'm finally getting more interested in the many alternative formulations and consequences, some of which sound more conceptual, like
If is a surjective endomorphism of a finitely generated -module, then is an isomorphism.
Yes, this is a good one. Nakayama's lemma seems to be like Yoneda's lemma, being one of these simple results that people in the know use all the time and is coincidentally named after a Japanese mathematician.
I'm not particularly in the know, but over a decade ago I wrote at the nLab, "Nakayama’s lemma is a simple but fundamental result of commutative algebra frequently used to lift information from the fiber of a sheaf over a point (as for example a coherent sheaf over a scheme) to give information on the stalk at that point".
I began writing the article after getting some glimmers of what it was really about (in the minds of people who really use the stuff) from this MathOverflow thread, I think especially the answer by Roy Smith. I especially like the application to the algebraic geometry form of the inverse function theorem.
James Deikun said:
The ideal does have to be finitely generated to apply Nakayama's Lemma, indeed.
Yes, I meant to say "finitely generated ideal".
Here's a nice counterexample when your ideal is not finitely generated. Let be the ring of continuous complex-valued functions on and let be the ideal consisting of functions that vanish at . Then but is not of the form for any idempotent (which would need to be a continuous function that only takes the values and/or ).
Todd Trimble said:
Nakayama's lemma seems to be like Yoneda's lemma, being one of these simple results that people in the know use all the time and is coincidentally named after a Japanese mathematician.
I was thinking about that. But what I'd really like to do is derive Nakayama's lemma from the Yoneda lemma! Then I could annoy people by saying "Nakayama's lemma? That's just a corollary of the Yoneda lemma."
Now, it may be hopeless to derive the Nakayama lemma from the Yoneda lemma. But the Nakayama lemma does follow from ideas connected to the Cayley-Hamilton theorem, and that theorem has a vaguely Yoneda-ish feel, at least in my fevered brain.
Remember, it says that if you take a square matrix and define its characteristic polynomial by
then you get
There's some sort of weird self-reference here, like " is the walking root of its own characteristic polynomial".
And that "walking" business feels vaguely Yoneda-esque (to my fevered brain).
Here's a fake one-line proof of the Cayley-Hamilton theorem:
Puzzle for beginners: spot why this proof is fake.
Puzzle for experts: try to morph it into a real proof!
John Baez said:
Puzzle for beginners: spot why this proof is fake.
Puzzle for experts: try to morph it into a real proof!
I'd be glad to see (or better find) a proof which use modern tools, ie. that I could understand easily (in my head, the more there is structure and the less there are tricks, the easier it is to understand). A modern definition of the determinant is like this:
For every -vector space of dimension , and for every , there is the exterior power of which is the quotient of the tensor power by the equalities for every permutation and where is the signature of (the next exterior powers are not very interesting because they are equal to )
If is a basis of , then the where form a basis of . It follows that . In particular is a basis of which is then of dimension .
As for tensor powers, you can take the exterior power of a map. If is an endomorphism, then sends to for a unique .
The definition of is then:
By definition, the characteristic polynomial of is then given by:
Let's verify that is really a polynomial (map) first.
is the map from to which acts like this on the basis
The matrix of is defined by the equations
Let me use into . Now I get that is the map from to which acts like this on the basis:
ie.
where is a uniquely determined scalar.
Remember that which gives in particular that when there is a couple such that .
For what it's worth, the Cayley-Hamilton theorem is proved in the nLab here, as a relatively easy consequence of Cramer's rule. By the way, Cayley-Hamilton holds over any commutative ring.
Yeah, thanks. I'm just trying to do my proof (I do the puzzle).
I prefer if there is no use of something like the Cramer's rule and if it follows from a direct computation by applying the definitions (that's probably a matter of taste). But I can't guarantee that it is not going to take me the whole day -- if it works.
So, I continue my computation.
Let me rewrite slightly differently what I wrote before (I deleted all the terms which are equal to using and enumerated the other ones using the symmetric group):
John Baez said:
Todd Trimble said:
Nakayama's lemma seems to be like Yoneda's lemma, being one of these simple results that people in the know use all the time and is coincidentally named after a Japanese mathematician.
I was thinking about that. But what I'd really like to do is derive Nakayama's lemma from the Yoneda lemma! Then I could annoy people by saying "Nakayama's lemma? That's just a corollary of the Yoneda lemma."
That's a fun motivation: finding a new way to annoy people using the Yoneda lemma. JK
I don't have a lot to say at the moment about a connection between the two, but it does remind me of stuff that Andre Scedrov said near the beginning of his AMS Memoir Forcing and Classifying Topoi. On page 11, he cites one of the standard approaches to proving the Cayley-Hamilton theorem as an example of universal thinking: "one always takes only the bare essentials needed to satisfy the hypothesis". So here, he says, "If ... , then showing at once is too hard. Rather, think of as a matrix over , as a module over , and show in this setting."
Where is he going with this? He's making an analogy with the process also described by Lawvere in his Variable Sets, Etendu, and Variable Structures in Topoi: just as one goes from a ring of constants to by adjoining a variable, so one can move from "constant sets" to more "variable sets" (e.g., presheaf toposes), and then the forcing process (localizing with respect to forcing conditions) is analogous to taking a quotient modulo an ideal, tailor-made to fit the hypotheses one is after.
So there is the idea of using generic elements. In the Yoneda lemma, one considers an identity morphism as a "generic generalized element".
That's all I got at the moment... wish I had more.
Jean-Baptiste Vienney said:
Yeah, thanks. I'm just trying to do my proof (I do the puzzle).
No, I know, and sorry to interrupt -- I thought you were done for the moment. Actually, I wasn't mainly talking to you, I was talking to everybody. Carry on...
Todd Trimble said:
Jean-Baptiste Vienney said:
Yeah, thanks. I'm just trying to do my proof (I do the puzzle).
No, I know, and sorry to interrupt -- I thought you were done for the moment. Actually, I wasn't mainly talking to you, I was talking to everybody. Carry on...
Ok :sweat_smile: . I wasn't sure if you were annoyed by my try or not. I take pauses.
Actually, I've never had a good understanding of this theorem and I didn't know exterior powers or even the tensor product when I learned it so I'm trying to see what I can do today using these tools (by the way I haven't used the hypothesis that the scalars are in a field, maybe the one that if , not sure)
I use again to get:
So, we have:
The are polynomials in (of degree ), which depend on the matrix of , although I don't want to write down the expression explicitely...
So is really a polynomial (of degree ) in .
Now, we want to prove that
Well, I can't make any further progress if I don't work on the coefficients
I can decompose
Jean-Baptiste Vienney said:
I'm gonna try to compute some coefficients here to get an idea
The coeff before is
The coeff before (if we had ) would be
The coeff before would be
So, there is a pattern
The coeff before is my where for every :
if
if
Let's combine :
Now, it's very clear that is a polynomial!
We thus want to prove that:
Be aware that the second product must be interpreted as a composite...
Let's note the dual basis of
I'm going to use the natural isomorphism
which sends an endomorphism to
(Note that is just the coefficients in the matrix of )
In fact it is also an isomorphism of -algebras
The product in is given by
(The idea is that absorbs the initial entry, is an intermediate output, absorbs this intermediate output, and then is a final output, that's why you get the coefficient )
I'm going to rewrite the LHS of as the -ary product applied to an element of ie. I will explicit an element such that if we consider the -ary product , then the LHS of is equal to
I do that because it feels too hard to compute the composite/product directly
but it will still replace by coefficients of its matrix + basis and dual basis elements in this LHS, so it should be better
Yeah, @James Deikun help me!! (or if you have something else in the same vein)
I changed the product because in you compose with
Let's see:
Under my isomorphism of -algebras, the fact that I want to prove is now equivalent to proving that this vector is equal to :
but recall that the are just the coefficients of the matrix of
so this vector is equal to:
Bceause everything is linear, I can move all the coefficients out of the product . Let's do it in several steps.
First, is equal to:
I'm going to put all the coefficients before vectors of the type together in the above expression. I obtain:
I just do a tiny simplification by replacing by when they are equal. I obtain:
Clearly, the thing to do now is to develop the big tensor product inside the
I want to apply a distributivity like this (where is a finite set):
Here is
is
is
I want to find, given a function , how I can write in a good way (with , , as above)
doesn't depend on here
so that
ie. by replacing and by their respective expression:
So that I get
I just replace the by its expression:
And I put this in to get:
Ouch
I recall that the Cayley-Hamilton theorem is equivalent to: for every , if is a basis of , and the matrix of then is equal to ...
I can move the first sum out of to get:
I still must use distributivities to move stuff outside of and finally be able to perform the multiplication
I must distribute here:
and here:
In the first expression, I must use the multinomial theorem:
image.png
This is the reference @Todd Trimble gave above. So now I'm thinking of as "the walking endomorphism" ring...
Sorry, you can discuss. I still hope that that my infinite calculation can terminate (if ever)
Jean-Baptiste Vienney said:
I prefer if there is no use of something like the Cramer's rule and if it follows from a direct computation by applying the definitions (that's probably a matter of taste). But I can't guarantee that it is not going to take me the whole day -- if it works.
With regard to Cramer's rule: it's an easy consequence of the fact that the determinant is an alternating multilinear map . In other words, a direct computation by applying the definition of determinant.
But the only reason for mentioning it is that it leads directly to the adjugate matrix of a matrix , viz. a matrix such that . The presence of such a matrix is the key thing used in the nLab to prove Cayley-Hamilton, which has a nice short conceptual proof (essentially the same proof as in Lang's Algebra, if I remember correctly).
It looks more intelligent. I'm still intrigued to know if my brute-force calculation can finish
I'm all for experimentation, but I think it would be better to work these things out by yourself and not go through a long public calculation. I think the main problem may be that not many people here would be willing to read through all of it, and besides, people were beginning to have a conversation about Nakayama's lemma.
Mhh I agree
I'll try to finish this outside of this conversation
Sorry for the spam with gigantic tensors
Paul Garrett has a proof of Cayley Hamilton at the end of his algebra book which I find pretty interesting.
Is there any good proof that doesn't make use of adjugates?
Patrick Nicodemus said:
Paul Garrett has a proof of Cayley Hamilton at the end of his algebra book which I find pretty interesting.
I think I would like it because he uses exterior powers, the adjugate, but with an endomorphism and not a matrix, and also modules over like mentionned above
James Deikun said:
Is there any good proof that doesn't make use of adjugates?
Adjugates can be motivated pretty well from a more abstract algebraic point of view, they don't have to be defined purely in terms of some formula that acts on a matrix. I can elaborate on this
Can you say more?
James Deikun said:
Is there any good proof that doesn't make use of adjugates?
There's a kind of "dirty" proof where one sees that diagonal matrices solve the characteristic polynomial equation (since for a polynomial is a diagonal matrix whose diagonal entries are , where are the entries for ). Since the determinant is invariant under conjugation, we can then extend Cayley-Hamilton to the set of all diagonalizable matrices. Then, since diagonalizable matrices are dense in the set of all matrices, and since the set of matrices that satisfy CH is a closed set (even in the Zariski topology), the Cayley-Hamilton holds for all matrices.
One has to think a bit carefully though how well this holds up for all commutative rings. I'm not sure that density assertion does.
Ack, diagonalizable matrices are not dense of course. I guess I had in mind the classical case where the commutative ring is a field, and apply the idea that the statement of CH is insensitive to whether we stick with that field or pass to an extension field, and use this to pass to an algebraic closure and argue there. So, yeah, it's looking dirtier by the moment.
I consider free modules of finite rank over a given ring .
Fix a rank 1 free module (not necessarily with a given choice of isomorphism with ) and let abbreviate the functor .
A perfect pairing is a map such that both ways of currying give isomorphisms, so and .
For any perfect pairing , and for an endomorphism of , construct a map
and by virtue of the natural isomorphism i can regard this as a map .
This map is called the adjoint of and is uniquely defined by . Similarly if is an endomorphism of its adjoint is an endomorphism of .
Now take to be a free module of rank , let
and take to be . The wedge product is a perfect pairing.
If is any endomorphism of then because the exterior algebra is a functor is an endomorphism of . So it has an adjoint which is an endomorphism of . I call this endomorphism the adjugate of .
Its defining property is that it satisfies
.
For any basis of there are standard bases for and . Moreover as soon as a basis is fixed is isomorphic to its dual space and so are and . I can describe the computation of the matrices arising from these but it will be a bit tedious, i will leave it as an exercise
it's helpful here to know that if is an endomorphism of then its determinant can be defined as the unique scalar in such that is equal to as endomorphisms of .
that's why the adjugate is closely connected to the determinant.
From this perspective is there a way to make the adjugate still work for modules that are not free? Apparently the Cayley-Hamilton theorem itself still works in this situation, just as long as the ring is commutative.
(This is much nicer than picking a basis and manipulating a bunch of formulas though.)
i don't know of a reformulation of the proof where you get it for free but it's a corollary. If is fg take a surjection , then lifts to such that the square commutes. then you apply the theorem to and use this to get it for
Todd Trimble said:
There's a kind of "dirty" proof where one sees that diagonal matrices solve the characteristic polynomial equation (since for a polynomial is a diagonal matrix whose diagonal entries are , where are the entries for ). Since the determinant is invariant under conjugation, we can then extend Cayley-Hamilton to the set of all diagonalizable matrices. Then, since diagonalizable matrices are dense in the set of all matrices, and since the set of matrices that satisfy CH is a closed set (even in the Zariski topology), the Cayley-Hamilton holds for all matrices.
I kind of like dirty arguments where you prove something on a dense set and then use continuity to handle the rest. Maybe it's my background in analysis: these arguments come very naturally to me. For example, that's one of my favorite proofs of
But you know this better than anyone, Todd - in our recent work I've been reaching for Zariski density arguments like a drunk reaches for his bottle. :upside_down:
I do however agree that there's something unsatisfying about proving an equation between algebraic functions by proving it at just enough points to let a density argument do the rest: in doing so one may be missing the deeper insights required to prove it more directly.
Todd Trimble said:
Ack, diagonalizable matrices are not dense of course. I guess I had in mind the classical case where the commutative ring is a field, and apply the idea that the statement of CH is insensitive to whether we stick with that field or pass to an extension field, and use this to pass to an algebraic closure and argue there. So, yeah, it's looking dirtier by the moment.
I'm wondering if there is an analogue of Smith normal form/Jordan canonical form for say PID entries (not like SNF, where you can pick different bases on either side, though), and then prove Cayley–Hamilton for that special form.
Or even just the pedestrian step of reducing to the case from a block diagonal matrix to the blocks separately, and then thinking about what happens with a single Jordan block, say (or the analogue not over a field).
Obviously when working over a more general commutative ring you almost surely have problems in finding invariant submodules and more options for what submodules can be. Checking now, I see that "hereditary" rings are those where submodules of free modules are projective, so things can get pretty weird, and this is probably not a path all the way to a general proof.
In general I don't find any of the proofs of Cayley-Hamilton very conceptual, and none of them seem to prove it in the full generality where it holds. In particular, any proof based on the adjugate doesn't seem like it would work for the case of quaternionic or split-quaternionic matrices. (How do you actually define determinant of quaternionic matrices anyway, BTW?)
Does Cayley-Hamilton even hold for quaternionic matrices?
I don't think there's a quaternion-valued determinant of quaternionic matrices obeying det(gh) = det(g) det(h). (It would take more conditions to turn this thought into a theorem.)
The good determinants for quaternionic matrices are the Dieudonne determinant and the Study determinant, which works out to be the square of the Dieudonne determinant. I wrote a bunch about these for the nLab.
The Dieudonne determinant is an idea that works for matrices valued in any division ring.
According to https://www.sciencedirect.com/science/article/pii/0024379595005439 it's a little more complicated than the paradigm case; the "characteristic polynomial" here detects right eigenvalues () and is obtained by intervening in the process of obtaining the Study determinant by doing the manipulation just before taking the complex determinant. It is a polynomial of degree with real coefficients, and its complex roots come in conjugate pairs, even the real ones, corresponding to "conjugate 3-spheres" of right eigenvalues, counting multiplicity. As a real polynomial, it is pretty clear how to evaluate it even on a quaternionic matrix and the evaluation on turns out to be 0.
@Jean-Baptiste Vienney pinging you as you requested this longer explanation.
Garibaldi in https://arxiv.org/pdf/math/0203276.pdf constructs the characteristic polynomial as basically the "stable minimal polynomial", encoding the linear dependencies among powers of that are not "coincidental". In this setting Cayley-Hamilton is basically trivial; the interesting thing is proving that the characteristic polynomial detects eigenvalues and obeys the conventional formula.
Here are string diagrams based on @Patrick Nicodemus's description of adjugates, in case this helps anyone else too.
adj1.png
adj2.png
adj3.png
Lovely! We can easily draw the identity defining the characteristic polynomial by an equality between 2 string diagrams too. so is the unique scalar such that:
Now, I'm wondering how you can draw the identity which is named "Cayley-Hamilton theorem"
And if we can somehow draw the proof with equalities of string diagrams
From the above-referenced Garibaldi:
Let be an -basis for . Let for (commuting) indeterminates, and let be the quotient field of . We call a generic element. The -span of is a subspace of , so it must be finite-dimensional over . Hence there is a nonzero monic polynomial in of smallest degree such that , called the minimal polynomial for over .
I like his general approach but I would really like it better if things like this were manifestly basis-independent ... I don't really know how to make a generic element of in a basis-independent way, when is a general associative algebra, though, much less one that will actually work in this kind of construction ...
What's ? A free -module? Is a field here?
is an -algebra, and is a field; he says most of this also works for a commutative ring, though quite some of his proofs would need to be revised.
I guess the interesting fact that should prove the existence of the characteristic polynomial is that the power maps on generate a finite-dimensional subspace of (the module of endomorphisms of as an affine scheme) as a module over the coordinate ring of when is finitely generated. But I'm not sure how to prove that; proving it by pigeonholing as above only works for individual points and then you have to do a bunch of other proofs to show the points act coherently together.
I think that searching for basis-free proofs is great but I also think that at some point you will run up against the wall that if a theorem only holds for finite dimensional vector spaces or finitely generated modules you will inevitably run into a point in the proof that invokes that hypothesis in some way or another. Maybe you're looking for some kind of HoTT style proof where you assume the propositional truncation of "there exists a basis of length ", I have no idea what the theory of such "finite dimensional" vector spaces is.
Hm, actually I guess is finitely generated over when is finitely generated over ... so the same sort of counting argument does work after all. Proving that the coefficients are actually polynomials and not rational functions seems more difficult though ... makes me wonder if it's even true for general algebras when, e.g., is a non-noetherian commutative ring ...
Patrick Nicodemus said:
I think that searching for basis-free proofs is great but I also think that at some point you will run up against the wall that if a theorem only holds for finite dimensional vector spaces or finitely generated modules you will inevitably run into a point in the proof that invokes that hypothesis in some way or another. Maybe you're looking for some kind of HoTT style proof where you assume the propositional truncation of "there exists a basis of length ", I have no idea what the theory of such "finite dimensional" vector spaces is.
Maybe such a thing will be possible when we will have a categorical characterization of the category of finite-dimensional vector spaces. André Kornell has found Axioms for the category of Hilbert spaces together with Chris Heunen, and then Axioms for the category of sets and relations, which share some similarities with the category of finite-dimensional vector spaces, because they are both compact-closed categories enriched over commutative monoids. He did a talk last time at uOttawa and said that axioms for (which is a category equivalent to ) look more difficult to find that for , which seems paradoxical. I wanted to tell him that I find it logical because the definition of a finite-dimensional vector space is longer than the definiton of a generic vector space and therefore a finite-dimensional vector space is something richer than a generic vector space, even if for instance, a non-degenerated -autonomous category of vector spaces, ie. a good category of infinite-dimensional vector spaces is a category even richer and more complicated than . I didn't dare telling him this because I was to shy, so I say it here.
Now, would axioms for allow to reason without basis? I guess yes but I'm not sure of that.
Looking at the axioms for Hilbert spaces, I don't think that such axioms for would not speak about basis, because possessing "dagger biproducts" is one of these axioms , so it looks like biproducts would play a role. And biproducts are a way to talk about direct sums and thus also basis.
Anyway, I can't answer the question before these axioms are found.
I like the characterization of finite-dimensional vector spaces as precisely the dualizable ones. Mike Shulman and Kate Ponto have a good paper on this
https://arxiv.org/abs/1107.6032
Although tbh my understanding stops at thinking of these as enabling you to do "feedback loops" or "fixed point" arguments, I don't understand how that works.
This paper definitely helped me to understand why the trace is important.
Thanks, I'll look at this!
When is finite-dimensional,the bilinear map sending , is an isomorphism. If you choose a basis for and use this to give a dual basis for , then for every , we can look at where sends the basis elements, and for any vector in , we can break into , apply to the basis elements and recombine them on the other side.
Every vector space has a map given by the evaluation map but only the finite dimensional vector spaces have a natural map , which is given by using the above isomorphism and sending to the identity map . Formally you can express this by saying that is right adjoint to in the one-object bicategory arising from the monoidal structure on . The evaluation map is the counit of the adjunction and the map I just described arising from the identity is the unit.
I think this recent paper gives on of the best proofs of the Cayley-Hamilton theorem: Hasse–Schmidt Derivations and Cayley–Hamilton Theorem for Exterior Algebras. Given that I've been working on categoryfying the notion of Hasse-Schmidt derivation (through something that I call higher-order differential categories), it may be possible that someday I will be able to categorify this paper too which is about a graded-commutative version of Hasse-Schmidt derivations, and thus maybe find a coordinate-free proof of the Cayley-Hamilton theorem. But I haven't even written my work on categorifying Hasse-Schmidt derivations, so please don't ask me to talk about this. I just say that maybe in one year or two, I could provide a coordinate-free proof of the Cayley-Hamilton theorem.
I'll check out that paper!
Jean-Baptiste Vienney said:
I think that you are right in looking at this with Grassmann's eyes... Then you should get that the coefficients of the determinant, gleaned from the n-fold exterior product should be the intrinsic "symmetric" funtions of an endomorphism (or matrix if you prefer), to wit, determinant, trace and the interpolations between them, essentially trace of the k-th exterior power, is that right?
The baby challenge would be to give a direct intrinsic proof of the fact that. (*). , where we got
as (the scale factor of the homothety) and as the scale factor of the homothety . Of course, you can prove (*) by brute force...
Todd Trimble said:
Jean-Baptiste Vienney said:
I prefer if there is no use of something like the Cramer's rule and if it follows from a direct computation by applying the definitions (that's probably a matter of taste). But I can't guarantee that it is not going to take me the whole day -- if it works.
With regard to Cramer's rule: it's an easy consequence of the fact that the determinant is an alternating multilinear map . In other words, a direct computation by applying the definition of determinant.
But the only reason for mentioning it is that it leads directly to the adjugate matrix of a matrix , viz. a matrix such that . The presence of such a matrix is the key thing used in the nLab to prove Cayley-Hamilton, which has a nice short conceptual proof (essentially the same proof as in Lang's Algebra, if I remember correctly).
I agree that the adjugate matrix (or endomorphism) of A is very helpful. Indeed, it is just the (n-1) exterior power of A...On the other hand, Cramer's are obvious if you look at your system of equations with Grassmann's eyes, taking advantage of the exterior product, if I remember well.
Jorge Soto-Andrade said:
Todd Trimble said:
Jean-Baptiste Vienney said:
I prefer if there is no use of something like the Cramer's rule and if it follows from a direct computation by applying the definitions (that's probably a matter of taste). But I can't guarantee that it is not going to take me the whole day -- if it works.
With regard to Cramer's rule: it's an easy consequence of the fact that the determinant is an alternating multilinear map . In other words, a direct computation by applying the definition of determinant.
But the only reason for mentioning it is that it leads directly to the adjugate matrix of a matrix , viz. a matrix such that . The presence of such a matrix is the key thing used in the nLab to prove Cayley-Hamilton, which has a nice short conceptual proof (essentially the same proof as in Lang's Algebra, if I remember correctly).
I agree that the adjugate matrix (or endomorphism) of A is very helpful. Indeed, it is just the (n-1) exterior power of A...On the other hand, Cramer's are obvious if you look at your system of equations with Grassmann's eyes, taking advantage of the exterior product, if I remember well.
This is an insightful comment, and fits together well with some other insightful comments (by Patrick Nicodemus, and by Aaron Fairbanks who produced those very helpful string diagrams), and here I am now catching up with all this.
Translating what Jorge is saying, the adjugate of an endomorphism for -dimensional is in fact the map defined as the composite
where the isomorphisms come from the perfect pairing . This definition of the adjugate was embodied in the string diagrams. By that definition, the desired equation is equivalent to commutativity of a pentagon diagram one of whose legs is
and the other of which is
Commutativity of this pentagon unwinds, string-diagrammatically, to a naturality square for which says that the composite
equals the composite
,
since by the conceptual definition of determinant.
Some interesting, and occasionally annoying, facts about commutative rings and their modules that I found out in the course of investigating all this:
I'd really like to find more connections between, and justifications for, these facts. I've seen proofs for them in at least some special cases, but they seem more like "coincidences of calculation" than proper justifications.
In particular, it's pretty surprising that the ideal is seemingly "almost always" generated by a single polynomial even when the ring is not a field; since a single-indeterminate polynomial ring being a principal ideal domain is equivalent to its coefficient ring being a field.
Here is a more basis-free version of the nLab proof of Cayley-Hamilton. (Were people able to follow that proof? I seemed to need to use the transpose of instead of .)
Let be a commutative ring, a free -module, and an endomorphism of .
Let denote the -module structure on induced by the unique -algebra map sending to . Let be the free-forgetful adjunction between and . Let denote the canonical map , which is the component of the adjunction counit at .
By the definition of , we have
.
Naturality of the counit then yields
,
i.e.,
.
Hence,
.
Wow, this looks elegant! I wish I understood it! I first run into trouble trying to understand what's . Maybe it should be obvious but what is this the identity morphism of?
I guess ?
Yes, that's right. Sorry, I should have somehow annotated that.
I think I get the proof now but it goes by so quickly it feels like I'm being fooled!
Probably part of the problem is that it's late and I'm tired. But if this proof is valid, it could really be the proof from The Book for the Cayley-Hamilton theorem!
(I never made it through the nLab proof.)
Wow, thanks! I appreciate the compliment. Let's hope we're not both being fooled, then.
Other folks should check out this proof! Like @Todd Trimble.
John Baez said:
(I never made it through the nLab proof.)
Oh for heaven's sake.
If only I hadn't encountered the idea of determinants and characteristic polynomials for non-matrix algebras I would probably consider this proof satisfying enough to foreclose my entire interest in the subject. But now I'm stuck pondering the True Meaning of the Determinant in general, and in particular the True Meaning of the Norm of the Sedonions. If it doesn't annihilate the zero divisors what good is it? Well, it's good for producing the characteristic polynomial! But why?
Todd Trimble said:
John Baez said:
(I never made it through the nLab proof.)
Oh for heaven's sake.
I'm doing lots of things. If I just glance at something out of random interest and it doesn't seem instantly delicious I'll switch to something else. Then there are things I'm actually working on, where I'm more persistent.
But sorry, I'd forgotten you were probably the one who wrote this proof. @Aaron David Fairbanks's proof might be a refashioning of the same proof - I don't know.
I haven't found any interesting applications of sedenions, or any deep connections between them and other topics in math or physics. I suspect they might be good for something, but I haven't found it.
This is quite different than the octonions, which have rich connections to exceptional Lie groups, superstring theory, etc.
The Wikipedia article says
Sedenion neural networks provide[ further explanation needed] a means of efficient and compact expression in machine learning applications and have been used in solving multiple time-series and traffic forecasting problems.
but I suspect this is basically bullshit, despite the 2 references.
So, right now I can't find "sedenion determinants" interesting.
I'm mostly interested in the sedenions here because they are an algebra where we both have a known expression for their "determinant" (the norm) that works for deriving the characteristic polynomial, and also it isn't a multiplicative homomorphism, making them a useful test case for figuring out what the true defining property of an algebra's determinant actually is.
Okay, this is a decent reason for being interested in them. Having spent years studying the octonions I keep hoping the sedenions will actually be connected to some deep mathematics, but so far I've seen no evidence of that. David Corfield calls
reals, complexes, quaternions, octonions, sedenions, ...
a 'sunken island chain', where the first islands poke above the water but the later ones don't.
John Baez said:
Todd Trimble said:
John Baez said:
(I never made it through the nLab proof.)
Oh for heaven's sake.
I'm doing lots of things. If I just glance at something out of random interest and it doesn't seem instantly delicious I'll switch to something else. Then there are things I'm actually working on, where I'm more persistent.
But sorry, I'd forgotten you were probably the one who wrote this proof. Aaron David Fairbanks's proof might be a refashioning of the same proof - I don't know.
I like Aaron's fashioning quite a bit -- it's nice that it's basis-free, and it is Book-like, and I'm considering editing the nLab proof as a result. But I don't think the proofs are all that different. For reference, I'm talking about the proof of Cayley-Hamilton, Theorem 1.2, which is only a few lines long. The equation on display in that proof is very close to the last line Aaron gave. In both cases, the proofs assume the stuff about the adjugate matrix as established beforehand; if the nLab stuff didn't have an inviting aroma, it may be because of the time spent establishing this first (Lemma 1.1).
Hi! Sorry for jumping in so late into this interesting discussion, but I just wanted to share a sort of divergent approach to Cayley-Hamilton I realized last week. First, I would rather begin from scratch, with a linear endomorphism of an n-dimensional vector space and the "natural" question: Can you find a "universal" polynomial equation, of degree not greater than , satisfied by ? It seems to me that you can get this polynomial "bare handed" just by some elementary combinatorial Yoga if you look at the whole situation with "Grassmann eyes".
Indeed, notice that for the baby case ( with ), we have
so that
in terms of the Grassmann-symmetric functions of .
Extension to general is rather straightforward. E.g. for , you begin with and end up analogously with
Notice that here by and so on, we mean the "pseudo endomorphism " (= span) of given by the morphisms and .
So, we play around with pseudo endomorphisms but we end up with a polynomial involving bona fide endomorphisms of .
Tod. d Trimble|277611** [said](https://categorytheory.zulipchat.com/#narrow/stream/266967-general.3A-mathematics/topic/Nakayama's.20lemma/near/389911724):
John Baez said:
Todd Trimble said:
John Baez said:
(I never made it through the nLab proof.)
Oh for heaven's sake.
I'm doing lots of things. If I just glance at something out of random interest and it doesn't seem instantly delicious I'll switch to something else. Then there are things I'm actually working on, where I'm more persistent.
But sorry, I'd forgotten you were probably the one who wrote this proof. Aaron David Fairbanks's proof might be a refashioning of the same proof - I don't know.
I like Aaron's fashioning quite a bit -- it's nice that it's basis-free, and it is Book-like, and I'm considering editing the nLab proof as a result. But I don't think the proofs are all that different. For reference, I'm talking about the proof of Cayley-Hamilton, Theorem 1.2, which is only a few lines long. The equation on display in that proof is very close to the last lin e Aaron gave. In both cases, the proofs assume the stuff about the adjugate matrix as established beforehand; if the nLab stuff didn't have an inviting aroma, it may be because of the time spent establishing this first (Lemma 1.1).
This looks really neat, but I guess I don't understand the general theory of these "pseudo endomorphisms" from your one example. What sort of maps am I allowed to use to form a pseudo endomorphism , and how is it defined?
John Baez said:
This looks really neat, but I guess I don't understand the general theory of these "pseudo endomorphisms" from your one example. What sort of maps am I allowed to use to form a pseudo endomorphism , and how is it defined?
Sorry, a bit more generally, for linear endomorphisms and from to the idea is to define as the span (pseudo morphism) from to itself consisting of the two linear maps from to given by and resp. In the approach I suggest, you only need the case .
You could define this more generally for any pair of linear maps , an also for n-tuples .
Of course, the naive definition does not work ( to fix that, some folks define
).
You seem to compose pseudo endomorphisms too, how does that work? The obvious thing would be the usual composition by pullback, but how does it look in Vect or wherever you're working?
James Deikun said:
You seem to compose pseudo endomorphisms too, how does that work? The obvious thing would be the usual composition by pullback, but how does it look in Vect or wherever you're working?
Yes, it looks like that, but more precisely, with the above notations, which is the pseudo endomorphism (span) defined by the maps and . A more general categorical setting is for sure possible ...
``` should be functorial on pairs of morphisms
I would guess that for this to work you just need a in your category which projects nicely onto the antisymmetric product. A key point though is that in Vect, for an n-dimensional vector space , the n-fold exterior product is isomorphic to the base field. A sort of "Grassmann - like" symmetric monoidal category...
Neat! By the way, the quote in your last comment has an error, so your reply to James is part of your quote of him. Your end-quote symbol ``` needs a blank line after it.
Patrick Nicodemus zei:
When is finite-dimensional,the bilinear map sending , is an isomorphism. If you choose a basis for and use this to give a dual basis for , then for every , we can look at where sends the basis elements, and for any vector in , we can break into , apply to the basis elements and recombine them on the other side.
Every vector space has a map given by the evaluation map but only the finite dimensional vector spaces have a natural map , which is given by using the above isomorphism and sending to the identity map . Formally you can express this by saying that is right adjoint to in the one-object bicategory arising from the monoidal structure on . The evaluation map is the counit of the adjunction and the map I just described arising from the identity is the unit.
Incidentally, this shows that Garibaldi’s generic element is uniquely determined: given an that occurs as the counit for an adjunction, the corresponding unit is unique, so if is the usual evaluation then Garibaldi’s is independent of the chosen basis.
John Baez said:
Neat! By the way, the quote in your last comment has an error, so your reply to James is part of your quote of him. Your end-quote symbol ``` needs a blank line after it.
Thanks for pointing this out.
By the way, I checked that the "natural" composition of my pseudo-endomorphisms coincide with the general composition of spans, via pullback (the pullback boiling down naturally to endowed with the "obvious" linear maps to . Notice however that we want not just any span from to itself as a result, but one which is an "avatar" of with left arrow being the "canonical" map (recall that "avatar" means "descent" ethymologically :blush: )