Professional Documents
Culture Documents
Series Editors
J. M. BALL W. T. GOWERS
N. J. HITCHIN L. NIRENBERG
R. PENROSE A. WILES
OX FO R D M AT H E M AT I C A L M O N O G R A P H S
DAVID A. BUCHSBAUM
Department of Mathematics, Brandeis University
1 3 5 7 9 10 8 6 4 2
A coloro che amo
From a little before the middle of the twentieth century, homological methods
have been applied to various parts of algebra (e.g. Lie Algebras, Associative
Algebras, Groups [finite and infinite]). In 1956, the book by H. Cartan and
S. Eilenberg, Homological Algebra [33], achieved a number of very important
results: it gave rise to the new discipline, Homological Algebra, it unified the
existing applications, and it indicated several directions for future study. Since
then, the number of developments and applications has grown beyond counting,
and there has, in some instances, even been enough time to see various methods
threading their way through apparently disparate, unrelated branches of algebra.
What we aim for in this book is to take a few homological themes (Koszul
complexes and their variations, resolutions in general) and show how these affect
the perception of certain problems in selected parts of algebra, as well as their
success in solving a number of them. The expectation is that an educated reader
will see connections between areas that he had not seen before, and will learn
techniques that will help in further research in these areas.
What we include will be discussed shortly in some detail; what we leave out
deserves some mention here. This is not a compendium of homological algebra,
nor is it a text on commutative algebra, combinatorics, or representation the-
ory; although, it makes significant contact with all of these fields. We are not
attempting to provide an encyclopedic work. As a result, we leave out vast areas
of these subjects and only select those parts that offer a coherence from the
point of view we are presenting. Even on that score we can make no claim to
completeness.
Our Chapter I, called “Recollections and Perspectives,” reviews parts of Poly-
nomial Ring and Power Series Ring Theory, Linear Algebra, and Multilinear
Algebra, and ties these with ideas that the reader should be very familiar with.
As the title of the chapter suggests, this is not a compendium of “assumed
known” items, but a presentation from a certain perspective—mainly homolo-
gical. For example, almost everyone knows about divisibility and factoriality; we
give a criterion for factoriality that ties it immediately to a homological inter-
pretation (and one which found significant application in solving a long-open
question in regular local ring theory).
The next three chapters of this book pull together a group of classical results,
all coming from and generalizing the techniques associated with the Koszul com-
plex. Perhaps the major result in Chapter II, on local rings, is the homological
characterization of a regular local ring by means of its global dimension. Section
II.6 includes a proof of the factoriality of regular local rings which is much closer
to the original one, rather than the Kaplansky proof that is frequently quoted.
viii Preface
I.1 Factorization
In this section, we deal with the basic topic of divisibility. In doing so, we review
a few properties of some rings, which are of importance to us. For more details,
we refer the reader to Reference [87].
exists for all choices of a and b in R − {0}, then lcm(a, b) exists for all choices
of a and b in R − {0} .
If 1 is a greatest common divisor for a and b, we will say that a and b are
coprime.
In terms of ideals, c being a common divisor of a and b means (a, b) ⊆ (c),
while c being a common multiple of a and b means (a) ∩ (b) ⊇ (c). The equality
(a, b) = (c) implies gcd(a, b) = c (but not conversely), while (a) ∩ (b) = (c) is
equivalent to lcm(a, b) = c.
reader is familiar with the process of forming rings of quotients with respect to
a multiplicative subset. Essentially this is just the “fractions” having arbitrary
elements of the ring on top, and elements of the multiplicative subset as denom-
inators. All the bells and whistles of localization are explained in Reference [41],
section 2.1). In fact, Rm is a PID (by the last proposition), because 0 and mm
are its only prime ideals. If mm = (a) for some a ∈ Rm , then every other ideal of
Rm is of type (an ) for some positive n.
Notice that since Rp is a PID for every p ∈ Spec(R), unique factorization of
elements is locally true for every Dedekind domain.
Local Dedekind domains are known as discrete valuation rings.
Although a similar theorem does not hold for R[[X1 , . . . , Xt ]], we have the
following partial result.
Proof We cannot reduce to the case t = 1. But since we already know that
R[[X1 , . . . , Xt ]] is a factorization domain, it suffices to show that every irreducible
element generates a (principal) prime ideal (cf. part (iii) of Proposition I.1.5).
This can be accomplished by induction on t, and using the statement of the
previous theorem, (cf., e.g. Reference [87], chapter 7, section 1, theorem 6, p. 148).
2
Proof The proof of part (ii) of Proposition I.1.13 works word for word in every
ring (R[[X1 , . . . , Xs−1 ]])[[Xs ]]. Since in our case R is a field, the non-units of
K[[X1 , . . . , Xt ]] are the elements with zero constant term. That is, (X1 , . . . , Xt )
consists of all the non-invertible elements of K[[X1 , . . . , Xt ]]. 2
We note immediately that the zero module is free (its basis is the empty
set). The following result shows that the basis of a free module can have any
cardinality.
Proposition I.2.2 Given any non-empty set I, there is a free R-module with
basis in one-to-one correspondence with I.
s
t
mi = aij mj , mj = bji mi .
j=1 i=1
Call S the t × s matrix (aij ) and T the s × t matrix (bji ). Clearly, ST equals the
t × t identity matrix It , and T S = Is . If t = s, say t > s, consider the square
matrices of order t:
T
S = (S | t − s zero columns) and T = .
t − s zero rows
We remark that the definition of determinant is the same as for fields, and
that det S det T = det ST is purely formal and does not require that the base
ring be a field.
Definition I.2.5 A free R-module is called finite if all its bases have finitely
many elements. A finite free R-module has rank t if it has a basis consisting of
t elements (hence every basis consists of t elements).
m1 f1
Proof (i) ⇒ (ii): S −1 ... = ... shows that C generates F (S −1
mt ft
exists for det S invertible allows the use of the customary formula); independence
is easy.
(ii) ⇒ (iii): Trivial.
(iii) ⇒ (i): Call T the matrix expressing B in terms of C; then T S = It , and
det S is a unit in R. 2
Corollary I.2.8
(i) Every generating system {m1 , . . . , mt } of Rt is a basis.
(ii) Every generating system of Rt has at least t elements.
(iii) If ϕ : Rn → Rm is an R-epimorphism, then n ≥ m.
Let T denote the t×t matrix ( It−s 0 | m1 · · · ms ). Since det T equals a maximal
minor of (m1 · · · ms ), a det T = 0. If T denotes the companion matrix of T ,
and Ti is the i-th column of T, then T T = det T · It implies:
1
..
. 0
1
T T1 · · · Tt−s aTt−s+1 Tt−s+2 · · · Tt = det T a ,
1
..
0 .
1
with a in position (t − s + 1, t − s + 1). It follows (since a det T = 0):
0
..
.
0
∗
0 = T aTt−s+1 = T
ab , with ab in row t − s + 1.
∗
.
..
∗
Hence 0 = abm1 + (∗m2 ) + · · · + (∗ms ), so that (since ab = 0) {m1 , . . . , ms }
cannot be an independent system.
Finally, we consider the case s > t, that is, (m1 · · · ms ) is a t × s matrix
with t < s. In that case, its maximal minors are of order t; if a kills all the
maximal minors, in particular it kills det(m1 · · · mt ). This implies (case s = t of
the induction hypothesis) that {m1 , . . . , mt } cannot be an independent system;
a fortiori, {m1 , . . . , ms } cannot be either.
This concludes the proof of the more general result on Rt , as well as of the
proposition. 2
Corollary I.2.10
(i) If ϕ : Rn → Rm is an R-monomorphism, then n ≤ m.
(ii) A non-zero ideal a of R is a finite free R-module if and only if it is
principal, and generated by a non-zero divisor.
ϕ
Proof (i) If n > m, then there would be a monomorphism Rn → Rm →
Rm ⊕ Rn−m = Rn associated with the n × n matrix S = (n−m zero S
rows) ,
where S is the matrix of ϕ with respect to the canonical bases. But det S = 0
contradicts the proposition.
(ii) If a is finite R-free, say a ∼
= Rt , the inclusion a → R gives a monomorphism
R → R, and t ≤ 1 by part (i). So a = Ra = (a) for some independent (that is,
t
equals β.
Conditions for being split are easily proven.
14 Recollections and perspectives
Proposition I.2.12 The following are equivalent for the exact sequence (∗∗).
The family of projective R-modules does not always properly include the
family of free R-modules.
Proposition I.2.17 Let (R, m) be a local ring (not necessarily noetherian) and
let M = 0 be a finitely generated R-module. If M is R-projective, then M is
R-free.
Proof Part (i) of the lemma implies that M/mM (= M ⊗R R/m) is a non-zero
finitely generated R/m-vector space, thus it has a basis, say {m1 , . . . , mt }. By
part (ii) of the lemma, it follows that {m1 , . . . , mt } is a generating system of M ,
hence there exists an epimorphism ϕ : F → M with F a rank t free module.
Since M is projective, ϕ splits and F = M ⊕ N for a suitable N . It follows that
F/mF = M/mM ⊕ N/mN as R/m-vector spaces. But F/mF and M/mM have
both dimension t, so that N/mN must have dimension 0, that is, N = mN and
N = 0 (again part (i) of the lemma). 2
The property of being projective is a local property, in the sense of the following
proposition.
Proposition I.2.18 Let M be a finitely presented R-module. M is projective
if and only if Mm is Rm -free for every m ∈ Max(R).
Proof The only if part follows from Proposition I.2.17, once we remark that
Mm = M ⊗R Rm R-projective implies Mm is Rm -projective (this comes from
Proposition I.2.15 (iii) and the fact that ⊗R commutes with ⊕).
The if part uses Proposition I.2.15 (iv), coupled with the remark that a map is
onto if and only if its localizations at all maximal ideals are onto. The assumption
that M is finitely presented ensures that
Rm ⊗R HomR (M, N ) ∼
= HomRm (Mm , Nm )
for every R-module N . 2
We now briefly turn our attention to a special class of rings.
Definition I.2.19 A nontrivial commutative ring (with 1) R is called hered-
itary if every ideal of R is a projective R-module.
Clearly every PID is hereditary, because of Corollary I.2.10 (ii). Another
example is given by Dedekind domains, due to the following result.
Proposition I.2.20 Every non-zero ideal a of a Dedekind domain R is a
projective R-module.
Proof Since R is noetherian, a is finitely presented as an R-module. Hence
by the last proposition, a is R-projective if and only if am is Rm -free for every
m ∈ Max(R).
Now given m ∈ Max(R), am ⊆ Rm satisfies am = (an ), where (a) is
the maximal ideal of Rm and n is a positive integer (recall the observa-
tions after Proposition I.1.12). Being a principal ideal, am is a rank 1 free
Rm -module. 2
Remark I.2.21 Dedekind domains are a more interesting example of hered-
itary rings than principal ideal domains, because in general not all ideals of
a Dedekind domain are free (if they were, they should all be principal, by
Remark I.2.6 and Corollary I.2.10 (ii)).
Theorem I.2.22 If R is hereditary, then every submodule of a free R-module
is R-isomorphic to a direct sum of ideals of R.
Proof Cf., for example, Reference [33], chapter I, theorem 5.3, p. 13. 2
Corollary I.2.23 If R is a PID, then every submodule of a free R-module is
free; in particular, every projective R-module P is free.
Corollary I.2.24 R is hereditary if and only if every submodule of a projective
R-module is projective.
Linear algebra 17
0 (M, N ) = M ⊗R N .)
(Notice that TorR
It is well known that the definition of TorR n (M, N ) does not depend on the
choice of the projective resolution of M . Furthermore, if one picks a projective
resolution of N and tensors it by M , the resulting truncated complex yields the
same homology modules. That is, TorR R
n (M, N ) = Torn (N, M ).
Clearly, Torn (M, N ) = 0 for every n ≥ 1, whenever either M or N is
R
projective.
As one might expect, Tor has some connection with the notion of torsion.
If N is an R-module, an element a ∈ N is said to be a torsion element
if there is a non-zero divisor r ∈ R such that ra = 0. The subset Ntor =
{a ∈ N | a is a torsion element} is clearly a submodule of N , and is called the
torsion submodule of N . N is called torsion-free if Ntor = 0; it is called a
torsion module if N = Ntor .
Example I.2.27 Let R be a commutative ring and r ∈ R a non-zero divisor.
Then
r
0 → R → R → R/(r) → 0
Linear algebra 19
is a projective (in fact free) resolution of R/(r), and for every R-module N , we
have that
It is well known that Tor “restores exactness” in the following sense. Given a
short exact sequence of R-modules:
0 → M → M → M → 0,
tensoring by N may destroy exactness on the left. But one can prove that there
is a long exact sequence:
...
→ TorR
n (M , N ) → Torn (M, N ) → Torn (M , N ) →
R R
...
→ TorR
1 (M , N ) → Tor1 (M, N ) → Tor1 (M , N ) →
R R
M ⊗R N → M ⊗R N → M ⊗R N → 0.
Proposition I.2.28 Let R be a PID and N a finitely generated R-module: N
is torsion free if and only if N is free.
Proof The if part is in the example above. Let us prove the only if part.
Thanks to Corollary I.2.23, Proposition I.2.18, and the fact that localization
preserves finite generation and torsion-freeness, we may assume that R is local,
hence a discrete valuation ring with maximal ideal (a), say.
If the R/(a)-vector space N ⊗R R/(a) has dimension t, by Lemma I.2.16 (ii)
we can find a rank t free R-module F such that
0→K→F →N →0
is exact (K simply stands for the kernel of the R-surjection F → N ). Tensoring
by R/(a), one gets the long exact sequence:
Definition I.2.30 For every n ≥ 0, the n-th extension module ExtnR (M, N )
is defined as follows:
ker(◦d1 ) if n = 0
n
ExtR (M, N ) =
ker(◦dn+1 )
im(◦dn ) if n ≥ 1.
(Notice that since Ext0R (M, N ) is the kernel of the map ◦d1 , and since
1 ◦d
0 → HomR (M, N ) → HomR (P0 , N ) −→ HomR (P1 , N )
As in the case of Tor, one can check that the definition of ExtnR (M, N ) does
not depend on the choice of the projective resolution of M . But if one wants to
get the same homology modules starting with a resolution of N , it is necessary
to pick an injective resolution of N and apply HomR (M, ).
It is clear that ExtnR (M, N ) = 0 for every n ≥ 1, whenever M is projective.
Multilinear algebra 21
Moreover, one can prove that given a short exact sequence of R-modules:
β
0 → M → M → M → 0,
α
0 → N → L → M → 0.
Since Ext1R (M, N ) = 0 by assumption, the long exact sequence of Ext yields the
surjectivity of
◦α
HomR (L, N ) → HomR (N, N ).
1⊗m↓ ↓m
m
A⊗A → A
1⊗u u⊗1
A⊗R → A⊗A R⊗A → A⊗A
∼
=↓ ↓m ∼
=↓ ↓m
1 1
A → A A → A
(we have omitted subscripts, and 1 stands for the identity on A).
m↓ ↓m
1
A → A
(here τ is the R-linear map sending a1 ⊗ a2 to a2 ⊗ a1 ).
By an ideal (right, left, two-sided) of the R-algebra A we always mean
an ideal (right, left, two-sided) of the ring A.
Multilinear algebra 23
everything works. 2
The last proposition says that our polynomial ring S identifies with the
symmetric algebra S(F ), in the sense of the following definition.
Definition I.3.5 Let M be an R-module. We call a symmetric algebra
on M any pair (S(M ), ϕ) satisfying the following requirements: S(M ) is an
R-algebra, ϕ : M → S(M ) is an R-morphism, the elements of ϕ(M ) commute
in S(M ), and for every R-map ψ : M → A, with A an R-algebra, such that
the elements of ψ(M ) commute in A, there exixts a unique R-algebra morphism
χ : S(M ) → A verifying ψ = χ ◦ ϕ.
It is clear that, if symmetric algebras on M do exist, there is a unique
isomorphism from one to the other which is the identity on M .
24 Recollections and perspectives
cA : A → A ⊗R A (comultiplication), εA : A → R (counit)
Multilinear algebra 25
1⊗c↑ ↑c
c
A⊗A ←− A
1⊗ε ε⊗1
A ⊗ R ←− A ⊗ A R ⊗ A ←− A ⊗ A
∼
=↑ ↑c ∼
=↑ ↑c
1 1
A ←− A A ←− A
(as before, we have omitted subscripts, and 1 stands for the identity on A).
It should be noticed that the diagrams occurring in the last definition are
obtained by reversing arrows in those describing algebra properties.
S(M ) is an R-coalgebra with respect to the following counit and
comultiplication.
ε is the identity on S0 (M ) = R and the zero map on Si (M ) for every i ≥ 1.
c is constructed from the diagonal mapping ∆ : m ∈ M → (m, m) ∈ M × M :
since ∆ is R-linear, a morphism S(M ) → S(M × M ) of R-algebras is induced;
but S(M × M ) ∼ = S(M ) ⊗R S(M ) as R-algebras, and c is defined to be the
composite S(M ) → S(M × M ) ∼ = S(M ) ⊗R S(M ).
Remark I.3.8 In the above, the following definition is understood: given two
R-algebras A and B, the R-algebra A ⊗R B is the R-module A ⊗R B with
multiplication defined by
(a1 ⊗ b1 )(a2 ⊗ b2 ) = a1 a2 ⊗ b1 b2
u ⊗u
and unit given by the composition RA −→ R ⊗R R A−→B A ⊗R B.
The fact, stated above, that S(M × M ) = ∼ S(M ) ⊗R S(M ) is easily seen if
one takes, instead of M × M the direct sum M ⊕ M . It is clear that M ⊕ M is
the degree one part of both S(M ⊕ M ) and of S(M ) ⊗R S(M ). The universality
property of the functor S(M ) does the rest.
As for an explicit description of c in the case of S(M ), since the isomorphism
S(M × M ) ∼ = S(M ) ⊗R S(M ) is induced by (m1 , m2 ) → m1 ⊗ 1 + 1 ⊗ m2 , and
c is a morphism of algebras, one gets
k
c(m1 · · · mk ) = (mi ⊗ 1 + 1 ⊗ mi ) = (mi1 · · · mih ) ⊗ (mj1 · · · mjk−h ),
i=1
26 Recollections and perspectives
the sum ranging over all pairs of strictly increasing sequences i1 < · · · < ih and
j1 < · · · < jk−h such that
{i1 , . . . , ih } ∪ {j1 , . . . , jk−h } = {1, . . . , k}
(including the cases in which one of the two sequences is empty).
One should remark that S(M ) is a graded coalgebra, in the sense that all
maps are compatible with the given gradings.
Remark I.3.9 Since the comultiplication of S(M ) is obtained from the diag-
onal map M → M × M , it is customary to call diagonal map the
comultiplication of any coalgebra, replacing the letter c by ∆.
In order to explain what a (graded) Hopf algebra is, we need the definition
of the product of coalgebras. Let A and B be two given R-coalgebras. The
R-coalgebra A ⊗R B is the R-module A ⊗R B with counit εA⊗R B (a ⊗ b) =
εA (a)εB (b) and with diagonal map
∆A⊗R B (a ⊗ b) = (ai ⊗ bj ) ⊗ (ai ⊗ bj ),
i,j
where i ai ⊗ = ∆A (a) and j bj ⊗ bj = ∆B (b).
ai
We also need the following: given R-coalgebras A and B, an R-coalgebra
homorphism ϕ : A → B is an R-module map such that the diagrams
∆ ε
A −→
A
A⊗A A −→
A
R
ϕ↓ ↓ϕ⊗ϕ ϕ↓ ↓1
ε
B −→
B∆
B⊗B B −→
B
R
are commutative.
Definition I.3.10 A Hopf algebra over R is an R-module A, which is both
an algebra and a coalgebra over R, satisfying the following properties:
(1) the multiplication m and the unit u are coalgebra morphisms;
(2) the diagonal map ∆ and the counit ε are algebra morphisms;
(3) there exixts an R-module map sA : A → A (called the antipode) such that
the diagrams
1⊗s s⊗1
A ⊗ A −→ A ⊗ A A ⊗ A −→ A ⊗ A
∆↑ ↓m ∆↑ ↓m
u◦ε u◦ε
A −→ A A −→ A
are commutative.
Multilinear algebra 27
m↓ ↓m
1
A −→ A
is commutative.
If B has the structure of a graded R-coalgebra, we say it is a cocommutative
graded R-coalgebra if the following diagram
τB,B
B ⊗ B ←− B ⊗ B
∆↑ ↑∆
1
B ←− B
is commutative.
Remark I.3.12 The last definition also has an impact on the construction
of the graded R-algebra A ⊗R B. The multiplication defined on A ⊗R B in
Remark I.3.8 is the composition
1A ⊗τ ⊗1B mA ⊗mB
A⊗B⊗A⊗B −→ A⊗A⊗B⊗B −→ A⊗B
(with τ (b, a)=(a, b)) and applies to non-graded algebras, as well as to graded
algebras concentrated in even degrees (such as the symmetric algebras). But in the
28 Recollections and perspectives
for every z ∈ Tn (E) and for every n ∈ N. If z ∈ Dn (E), then sym(z) = n!z.
Let D(E) stand for the graded R-module ⊕n∈N Dn (E).
Proposition I.3.13 D(E) is a graded R-algebra.
Proof
Given z ∈ Di (E) and z ∈ Dj (E), then zz ∈ Di+j (E) is defined to be
σ∈Si,j σ(z ⊗ z ), where
Si,j = {σ ∈ Si+j | σ(1) < · · · < σ(i) and σ(i + 1) < · · · < σ(i + j)} .
The remainder of the proof is standard. 2
Remark I.3.14 If e1 , . . . , en belong to E, then their product e1 · · · en in D(E)
equals sym(e1 ⊗ · · · ⊗ en ).
Definition I.3.15 D(E) is called the divided power algebra on the free
R-module E.
If we assume that the elements of Dn (E) have degree 2n, then D(E) is
commutative as a graded R-algebra.
Multilinear algebra 29
where the sum ranges over all pairs (u1 , . . . , ut ) and (v1 , . . . , vt ) of elements of
Nt , such that ui + vi = si for every i in {1, . . . , t}. We leave all the details to the
reader.
The indicated basis of D(E) leads one to think that D(E) is a symmetric
algebra (a polynomial ring). This is indeed the case if R contains a copy of the
(s ) (s )
rationals. For then every basis element y1 1 · · · yt t can be replaced by
(s1 ) (st )
s1 ! · · · st !y1 · · · yt = y1s1 · · · ytst
and D(E) ∼ = R[y1 , . . . , yt ].
In general, however, D(E) is not a symmetric algebra. What is true, though,
is that D(E) is the graded dual of a symmetric algebra.
Remark I.3.18 It is well known that, given R-modules Mi , i ∈ N, and L,
HomR Mi , L ∼= HomR (Mi , L).
i∈N i∈N
In particular, the ordinary dual of a graded R-algebra A = ⊕i∈N Ai is i∈N A∗i .
Hence, in order to make sure that the dual of a graded algebra is still a graded
algebra, one replaces A∗ with another object, namely ⊕i∈N A∗i , which is called
the graded dual of A and is denoted by A∗gr .
30 Recollections and perspectives
Theorem I.3.19 D(E) is the graded dual of the symmetric algebra S(F ), and
S(F ) is the graded dual of the divided power algebra D(E).
If one uses the notation F ∗ in place of E, the above says that D(F ∗ ) = S(F )∗gr
and S(F ) = D(F ∗ )∗gr .
We sketch a proof of the theorem, by giving an idea of why D(E)∗gr equals
S(F ). Essentially, it all amounts to showing that the diagonal map of D(E)
induces on D(E)∗gr a graded R-algebra structure corresponding to that of S(F ).
Let us take f ∈ Di (E)∗ and g ∈ Dj (E)∗ , and use the diagonal map of
D(E) to define an element f g ∈ Di+j (E)∗ . Since f ∈ HomR (Di (E), R) and
g ∈ HomR (Dj (E), R), we may set f g equal to the composite map
∆ f ⊗g
Di+j (E) −→ Di (E) ⊗R Dj (E) −→ R ⊗R R −→ R,
an element of HomR (Di+j (E), R), as required. One checks that this multiplica-
tion gives D(E)∗gr (which equals S(F ) as an R-module) an R-algebra structure.
We claim that such a structure is the ordinary multiplication in S(F ).
Clearly, it suffices to prove that
∗ ∗ ∗
(s ) (s ) (r ) (r ) (s +r ) (s +r )
(∗) y1 1 · · · y t t y 1 1 · · · yt t = y1 1 1 · · · yt t t ,
∗
stands for the element of (Ds1 +···+st E)∗ = Ss1 +···+st F
(s1 ) (st )
where y1 · · · yt
(s ) (s )
dual to y1 1 · · · yt t (and due to correspond to xs11 · · · xst t ). But (∗) is obvious,
in view of the definition of ∆ in D(E) recorded above.
(2)
We must prove that the composite map assigns 1 to y1 y2 and 0 to all other
basis elements.
The action of y1∗ ⊗(y1 y2 )∗ on ∆(y1 y2 ) = y1 ⊗y1 y2 +y2 ⊗y1 yields 1⊗1+0⊗0,
(2) (2)
hence 1.
The action of y1∗ ⊗(y1 y ∗ (2) (2) (2)
2 ) on ∆(y
(3)
) = y1 ⊗ y1 , on ∆(y1 y2 ) = y1 ⊗ y2 +
(3) (2)
y2 ⊗ y1 y2 , and on ∆ y2 = y2 ⊗ y 2 yields (respectively) 1 ⊗ 0, 1 ⊗ 0 + 0 ⊗ 1,
and 0 ⊗ 0; hence 0 all the times.
Concerning the example, one should notice that the product of y1 and y1 y2 in
(2)
D(E) equals 2y1 y2 (by part 1 of Remark I.3.17), while x1 times x1 x2 in S(F )
gives x21 x2 .
where s stands for the number of ordered pairs (u, v) verifying jv < iu , and
the sum ranges over all pairs of strictly increasing sequences i1 < · · · < ih and
j1 < · · · < jk−h such that
{i1 , . . . , ih } ∪ {j1 , . . . , jk−h } = {1, . . . , k}
(including the cases in which one of the two sequences is empty).
We now focus our attention on Λ(F ), where F is a finite free R-module,
and reobtain some results of Subsection I.2.1, namely Propositions I.2.4, I.2.7,
and I.2.9.
In order to state the next lemma, we need the notion of an alternating n-linear
function on M . It is an n-linear function ψ : M × · · · × M → N , with N any
R-module, such that ψ(m1 , . . . , mn ) = 0 whenever two of the arguments coincide.
Clearly,
ϕ : M × · · · × M → Λn (M ), (m1 , . . . , mn ) → m1 ∧ · · · ∧ mn
is an alternating n-linear function. It is not hard to see that every ψ as above
factors through ϕ, that is, ψ = χ ◦ ϕ for some linear map χ : Λn (M ) → N .
Hence
t
ψ(m1 , . . . , mt )m = i , . . . , mt )mi .
(−1)i+1 ψ(m, m1 , . . . , m
i=1
simply equals the exterior product of the chosen summand and fI . Since it is
still an element of the two-sided ideal ker(Λ(ϕ)), if we apply Λ(ϕ) to it, we get:
0 = ai1 ,...,is ϕ(fi1 ) ∧ · · · ∧ ϕ(fis ) ∧ ϕ(fj1 ) ∧ · · · ∧ ϕ(fjk ).
0
Remark I.3.32
1. Let F be R-free with basis {f1 , . . . , ft }. Let {g1 , . . . , gt } be the basis of F ∗
dual to {f1 , . . . , ft }. Then for every k ∈ {1, · · · t}, we can define an isomorphism
Λk (F ∗ ) → (Λk (F ))∗ by sending each basis element gi1 ∧· · ·∧gik to the application
. , ik ) ∈ HomR (Λk (F ), R) defined by means of ϕ(i1 , . . . , ik )(fj1 ∧ · · · ∧
ϕ(i1 , . .
1 if ih = jh ∀h
fjk ) = . The indicated isomorphism is independent of the choice
0 otherwise
of the basis {f1 , . . . , ft }.
2. Let F be a rank t free R-module. If a non-zero element of Λt (F ) is chosen,
for example, an isomorphism ψ0 : Λt (F ) → R is fixed, then further isomorphisms
ψk : Λt−k (F ) → Λk (F ∗ ) are obtained for k ∈ {1, . . . , t}. Namely, given z ∈
Λt−k (F ) and w ∈ Λk (F ), one sets ψk (z)(w) = ψ0 (z ∧ w). This means that an
isomorphism Λ(F ) ∼ = Λ(F ∗ ) also exists. However, it is not independent of the
choice of the basis in F (= the choice of the element of Λt (F )).
Also the minors of a matrix can be described in terms of exterior multiplic-
ation. For let E and F be finite free R-modules, with bases {e1 , . . . , es } and
{f1 , . . . , ft }, respectively. Let ψ ∈ HomR (E, F ) have t × s matrix S with
respect
to the indicated bases, and let Λk (ψ), 1 ≤ k ≤ min{t, s}, have kt × ks matrix
U relative to the corresponding bases of Λk (E) and Λk (F ). Then the entry of
U , which is located at the intersection of the row indexed by the basis element
fi1 ∧ · · · ∧ fik and the column indexed by the basis element ej1 ∧ · · · ∧ ejk , is
precisely the k × k minor of S with rows indexed by fi1 , . . . , fik and columns
indexed by ej1 , . . . , ejk . Such a minor is often denoted by (i1 , . . . , ik | j1 , . . . , jk ).
It is customary to denote by Ik (ψ) the ideal generated by the entries of U ,
that is, by the k × k minors of S. The ideal Ik (ψ) is independent of the choice
of bases because the isomorphism
HomR (Λk (E), Λk (F )) ∼
= (Λk (E))∗ ⊗R Λk (F ),
coupled with
(Λk (E))∗ ⊗R Λk (F ) ∼
= (Λk (E) ⊗R Λk (F ∗ ))∗ = HomR (Λk (E) ⊗R Λk (F ∗ ), R),
identifies Λk (ψ) with a map Λk (E) ⊗R Λk (F ∗ ) → R, whose image is pre-
cisely Ik (ψ).
This suggests that we define
Λk (ψ)
Ik (ψ) = im(Λk (E) ⊗R Λk (F ∗ ) −→ R) ∀k ≥ 0,
which, among other things, implies that the ideal is 0 for k > min{t, s} (no minors
exist), and that the 0 × 0 minors generate R (a 0 × 0 matrix has determinant 1).
II
LOCAL RING THEORY
The development of the first five sections of this chapter is guided by the desire to
establish a string of inequalities for local rings (always assumed to be noetherian
here) which lead to the result, in Section II.5, that a local ring is regular if and
only if its global dimension is finite. The development here is very close to that
used in the notes of Reference [19] and rests on the work found in References [7]
and [8].
In the first section we define and develop the properties of the Koszul complex,
which will play a fundamental role throughout the book. The next section, on
local rings, ends with a statement of a fundamental inequality whose proof rests
almost completely on these properties. This inequality states that the global
dimension of a local ring always exceeds the minimal number of generators of
its maximal ideal. While Section II.3 on Hilbert–Samuel polynomials does not
use the complex itself, it does rely on the notion of “regular sequence” for some
of its fundamental results, and this notion is intimately connected with that
tool. In that section we establish the fact that the dimension of any local ring
is always exceeded by the minimum number of generators of its maximal ideal
(Krull’s Principal Ideal Theorem). The results of Section II.4 on codimension
and finitistic global dimension are inextricably tied up with the Koszul complex;
there we establish the facts that the finitistic global dimension of a local ring
is equal to its codimension, and that its codimension is never greater than its
dimension. Section II.5 on regular local rings puts together all of the above to
achieve the result mentioned in our first paragraph. Another main result there
is the Cohen–Macaulay Theorem; its main new “homological” point is that the
“unmixedness” part of the classical theorem is sharpened by the observation that
the Koszul complex associated to the generators of the ideal is acyclic, which then
yields the result.
The sixth section of this chapter contains a full proof of the unique factoriz-
ation theorem for regular local rings. This theorem, too, was one of the prime
objectives of the application of homological methods to commutative ring theory.
In the penultimate section, we take up the notion of multiplicity, which is again
intimately tied up with the Koszul complex. While a detailed study of this the-
ory is beyond the scope of our book, it would be remiss to completely ignore it.
In this section, therefore, we give a definition of multiplicity and establish some
of its properties by means of the Koszul complex. Connected with this topic is
the notion of Cohen–Macaulay ring, and we will see that for such rings the two
38 Local ring theory
invariants of an ideal, its height and its depth, are equal. As the depth of an
ideal is a number that in a certain sense (that of relative primeness) measures
the “arithmetic” size of an ideal, while the height is a number that measures
its “geometric” size, their coincidence in the case of Cohen–Macaulay rings sug-
gests why the use of ideal-theoretic methods in algebraic geometry, when the
variety is Cohen–Macaulay, is so effective.
From multiplicity in Section II.7, we move to the Serre definition of intersection
multiplicity in Section II.8 and we close the chapter with a description of the so-
called homological conjectures. This book does not deal with these conjectures,
but we feel that no book that purports to present the interaction of local ring
theory and homology can omit at least a brief mention of these fundamental
questions and the work that has been done on them.
where g(r1 , r2 ) = r1 x + r2 y and f (r) = (−ry, rx). Calling this complex K(x, y),
we have H0 (K(x, y)) = R/(x, y); H2 (K(x, y)) = 0 : (x, y) = {r ∈ R| rx = 0 =
ry}. H1 (K(x, y)) = {(r1 , r2 )| r1 x+r2 y = 0}/{(−ry, rx)}. If we suppose that R is
an integral domain, and that neither x nor y is zero, we see that H2 (K(x, y)) = 0
and that {(r1 , r2 )| r1 x + r2 y = 0} is mapped isomorphically onto (x) : y = {r ∈
R| ry ∈ (x)} by the map sending (r1 , r2 ) to r2 . Under this map, the submodule
{(−ry, rx)} is sent onto the ideal (x) and thus H1 (K(x, y)) = (x) : y/(x). If R
were a unique factorization domain, then H1 (K(x, y)) would be zero if and only
if x and y were relatively prime. Therefore, we again see that a certain amount
of homological information is intimately connected with arithmetical questions.
We could proceed step by step in this way, building up complexes
K(x, y, z, . . .), but it is more efficient to consider the following general construc-
tion. Let f : A → R be a morphism from the R-module A to R. Then we
may define the Koszul complex, K(f ), by setting K(f )p = Λp A and defining
df : Λp A → Λp−1 A by
∧
df (a1 ∧ · · · ∧ ap ) = (−1)i+1 f (ai ) a1 ∧ · · · ∧ ai ∧ · · · ∧ ap ,
∧
where ai means we omit ai . It is well known, and easy to check, that (df )2 = 0.
Also, if α ∈ Λp A and β ∈ Λq A, then
To see how this subsumes the sort of thing we have been discussing, consider
a sequence, x1 , . . . , xs , of elements of our ring R. We can define a morphism
f : Rs → R, where Rs denotes the direct sum of R with itself s times, and f is
defined by f (1, . . . , 0) = x1 , . . . , f (0, 0, . . . , 1) = xs . In this case, we denote the
complex K(f ) by K(x1 , . . . , xs ). For s = 1 or s = 2, we obtain the complexes
K(x) and K(x, y) described above.
Before we return to the general case of a morphism f : A → R, let us review the
mapping cone construction for a map between complexes. First, a convention
about complexes. Since in most of our applications our complexes will be left
complexes, we will assume that a complex C is given by {Cn ; dn } with the
boundary map d = {dn } of degree −1. That is, for each n, dn : Cn → Cn−1 , and
we will usually write the complex:
d d d d d−1 d−2
· · · →3 C2 →2 C1 →1 C0 →0 C−1 → C−2 → · · · .
Thus, our complex K(f ) above has all its negative parts equal to zero, K(f )0 =
R, K(f )1 = A, and so on.
M (F )n = Dn ⊕ Cn−1 ;
40 Local ring theory
(Here we are writing the boundary map for the complex D as d = {dn }.)
Notice that, module-theoretically, the complex M(F ) is the direct sum of the
complexes C and D, except for one thing: the degree of the component coming
from C is off by one. If we denote by C the complex {C n ; dn }, where C n = Cn−1
and dn = dn−1 , then our mapping cone is, in each degree, the direct sum of C
and D.
Proposition II.1.2 Let F : C → D be a map of complexes. Then we have the
short exact sequence of complexes:
(∗) 0 → D → M(F ) → C → 0
and the long exact sequence of homology:
· · · → Hn+1 (C) → Hn (D) → Hn (M(F )) → Hn (C) → · · ·
where the maps Hn+1 (C) → Hn (D) are the maps in homology induced by ±F.
Proof The exactness of the short exact sequence (∗) is clear on the module-
theoretic level. What has to be checked is that the obvious injection and
projection maps are maps of complexes. But this is quite simple. Once the
exactness of (∗) is established, the long exact homology sequence is an imme-
diate consequence; what must be explained is the remark that the maps
Hn+1 (C) → Hn (D) are the maps in homology induced by ±F. From the defini-
tion of C, it is clear that Hn+1 (C) = Hn (C), so that Hn+1 (C) → Hn (D) is really
Hn (C) → Hn (D). But the definition of the boundary map of M(F ) involves
the map F with a sign. A careful diagram chase to describe the “connecting
homorphism” makes it clear that this morphism is induced by the map F in
homology, with the sign determined in each dimension according to the definition
of δn . 2
Returning to the general case of a morphism f : A → R, we state the funda-
mental properties of the complex K(f ) :
Proposition II.1.3 If a0 is an element of A, then f (a0 )Hp (K(f )) = 0 for all
p ≥ 0.
Proposition II.1.4 If x is any element of R, and f : A → R a morphism, we
have the map of complexes x : K(f ) → K(f ) :
f
· · · → Λ3 A → Λ2 A → A → R
↓x ↓x ↓x ↓x .
f
· · · → Λ3 A → Λ 2 A → A → R
Koszul complexes 41
The mapping cone of this map of complexes is denoted by K(f ) where f : A⊕R →
R is defined by f (a, r) = f (a) + rx.
for all p. Proposition II.1.4 is proved by observing that for any module A, Λp (A⊕
R) ∼= Λp A ⊕ Λp−1 A, and then checking that under this identification the map
df is the boundary map of the mapping cone. 2
Proof All but the last statement about the exactness of (E(M )) follows
immediately from the propositions cited. The exactness of (E(M )) is due
to the fact that the short exact sequence (∗) applied to the complexes
C = D =K(x1 , . . . , xs ) and K(x1 , . . . , xs+1 ) = M(F ), where F is multiplica-
tion by xs+1 , splits. Thus tensoring by M preserves exactness, and the long
homology sequence (E(M )) is the one associated to the short one. 2
From the remarks made about the meaning of H1 (K(x, y)), we are led to make
the following definition.
42 Local ring theory
elements. This number is equal to the dimension of the vector space E/mE over
k, and is denoted by [E/mE : k]. Moreover, any subset of E linearly independent
modulo mE may be extended to a minimal generating set of E.
Lemma II.2.2 If E is an R-module, there exists a free module F and an
epimorphism g : F → E such that ker(g) is contained in mF.
Proof The proof depends on nothing deeper than the fact that an epimorphism
of a vector space onto another of the same dimension is an isomorphism. One
then chooses F to be free on the number of generators in a minimal generating
set of E. 2
Lemma II.2.3 If E is an R-module such that TorR
1 (k, E) = 0, then E is free.
Since the map F/mF → E/mE is an isomorphism, the map TorR 1 (k, E) →
K/mK is surjective. But here we can invoke the fact that TorR
1 (k, E) = 0 to
conclude that K/mK = 0 and, by Nakayama, K itself must be zero. 2
Lemma II.2.4 For every R-module E, hdR (E) ≤ hdR (k).
Lemma II.2.5 If R is a local ring, gldimR = hdR k.
Recall that the global dimension of R, gldimR, is defined to be sup hdR (M )
M
as M runs through all R-modules.
Lemmas II.2.4 and II.2.5 are easy consequences of Lemma II.2.3.
The main result we are heading for now is that [m/m2 : k] ≤ gldimR, that
is, that the global dimension of R is never less than the minimal number of
generators of the maximal ideal of R. The idea of the proof is the following.
We take a minimal generating set x1 , . . . , xn of the maximal ideal m, and form
the complex K(x1 , . . . , xn ). We then show that we can find a free resolution of k :
d d d
· · · → X3 →3 X2 →2 X1 →1 R → k → 0
such that di (Xi ) ⊂ mXi−1 and such that K(x1 , . . . , xn ) is a subcomplex of
this resolution. Tensoring this resolution with k makes the boundary maps
zero, so that TorR p (k, k) = Xp /mXp . Since K(x1 , . . . , xn ) is a subcomplex
of the resolution, and K(x1 , . . . , xn ) is non-zero in dimension n, we see that
TorRn (k, k) = 0. Thus hdR k ≥ n and we have the desired result using the facts
that gldimR = hdR k and n = [m/m2 : k].
Local rings 45
d d
Lemma II.3.3 If f (n) = i=0 ni ai = i=0 ni ai is a polynomial function,
then d = d and ai = ai for all i. Thus, the degree, d, of f is a well-defined
integer, and the coefficients ai of f are uniquely determined.
generated graded R[X1 , . . . , Xs ]-modules E = ν≥0 Eν such that
1. if s = 0, Eν is in A for all ν;
2. if s> 0, Eν is in A for all ν and the graded R[X1 , . . . , Xs−1]-modules
Xs Xs
ker ν≥0 Eν → ν≥0 Eν and coker ν≥0 Eν → ν≥0 Eν are in
As−1 .
Finally, let f0 be a function from the objects of A to an abelian group G which
factors through the Grothendieck group, K(A), of A, that is, f0 is additive with
respect to exact sequences in A. Then
Theorem II.3.4 Let E = ν≥0 Eν be an object in As and define fE : Z+ → G
by fE (ν) = f0 (Eν ). Then fE is a polynomial function of degree less than or equal
to s − 1.
An example of such a set-up occurs when R is a local ring, and A is the full
subcategory
of R-modules of finite length. In that case, an R[X1 , . . . , Xs ]-module
E = ν≥0 Eν is in As if it is finitely generated, and if each Eν is an R-module
of finite length. We may then choose G to be the group, Z, of integers and f0 (E)
equals the length of E. This is of course the most usual example. To see that
it is not the only example, we may choose A to be the category of all finitely
generated R-modules, and G to be K(A) itself with f0 the usual map.
Corollary II.3.5 If R is a local ring, E a finitely generated R-module and q
an ideal of R containing some power mn of the maximal ideal m, then E/qν E
is an R-module of finite length, and the function χq (E) : Z+ → Z defined by
χq (E; ν) = length(E/qν E) is a polynomial function whose degree is less than or
equal to [q/mq : k].
To see why this is a corollary of Theorem II.3.4, we observe that we have exact
sequences
0 → qν E/qν+1 E → E/qν+1 E → E/qν E → 0
and therefore ∆χq (E; ν) equals the length of qν E/qν+1 E. The graded mod-
ule qν E/qν+1 E is a module over R[X1 , . . . , Xs ] where s = [q/mq : k], and
each q E/qν+1 E is an R-module of finite length. Thus ∆χq (E) is a polynomial
ν
Proposition II.3.8 Let R be a local ring, and let s be the smallest number of
elements required to generate an ideal q of R which contains some power of the
2
maximal ideal m. Then dim R ≤ s. In particular, dim R ≤ [m/m : k].
(I n M ) ∩ M = I n−h (I h M ∩ M ).
The proof depends on the Artin–Rees Theorem and the easy observation that
x(mν E : x) = mν E ∩ (x)E.
Hilbert–Samuel polynomials 49
Remark II.4.7
(1) For any R-module E, the set of zero divisors of E is equal to the union of
all primes in Ass(E).
(2) Thus, if E is a finitely generated R-module (with R noetherian) and if I
is an ideal such that every element of I is a zero divisor of E, then there is a
non-zero element e of E such that Ie = 0. For if I is as above, then I ⊂ ∪p where
p runs through Ass(E) and hence I ⊂ p for some p ∈ Ass(E) (since Ass(E) is
finite). We therefore have the epimorphism R/I → R/p and a monomorphism
R/p → E. The image of 1 under the composite map R → R/I → R/p → E is
the desired element e ∈ E.
(3) If the ideal I in 2 above is a maximal ideal, then I ∈ Ass(E).
We are now ready to prove the main theorem of this section.
Theorem II.4.8 Let R be a noetherian ring, E an R-module, and I an ideal
of R generated by elements x1 , . . . , xn such that E/IE = 0. Let y1 , . . . , ys be a
maximal E-sequence in I. Then s+q = n, where q is the dimension of the highest
non-vanishing homology of the complex E(x1 , . . . , xn ). Furthermore,
Hq (E(x1 , . . . , xn )) ∼
= ((y1 , . . . , ys )E : I)/(y1 , . . . , ys )E.
Proof The proof proceeds by induction on s. When s = 0, it means that every
element of I is a zero divisor for E so that by 2 of Remark II.4.7, there is a non-
zero element e in E such that Ie = 0. Since Hn (E(x1 , . . . , xn )) = 0 : I = 0, we
see that q = n and Hn (E(x1 , . . . , xn )) = (y1 , . . . , ys )E : I/(y1 , . . . , ys )E = 0 : I.
When s > 0, we consider the exact sequence
y1
0→E→E→E→0
and get the long exact sequence
y1
Hq+1 (E(x1 , . . . , xn )) → Hq (E(x1 , . . . , xn )) → Hq (E(x1 , . . . , xn ))
→ Hq (E(x1 , . . . , xn )) → · · ·
where q is the dimension of the highest non-vanishing homology group of
E(x1 , . . . , xn ). Thus, Hq+1 (E(x1 , . . . , xn )) = 0, Hq (E(x1 , . . . , xn )) = 0, and
multiplication by y1 on Hq (E(x1 , . . . , xn )) and Hq−1 (E(x1 , . . . , xn )) is zero
since y1 is in I. Thus Hq (E(x1 , . . . , xn )) = 0 while Hq−1 (E(x1 , . . . , xn )) ∼ =
Hq (E(x1 , . . . , xn )). Noting that y2 , . . . , ys is a maximal E-sequence in I, and
using induction, we have (s−1) +q = n. Since, however, we have just shown that
q = q − 1, we have s + q = n. Finally, since Hq (E(x1 , . . . , xn )) ∼ = (y2 , . . . , ys )E :
I/(y2 , . . . , ys )E ∼= (y1 , . . . , ys )E : I/(y1 , . . . , ys )E, and since, in addition,
Hq (E(x1 , . . . , xn )) = Hq−1 (E(x1 , . . . , xn )) ∼ = Hq (E(x1 , . . . , xn )), we have
∼
Hq (E(x1 , . . . , xn )) = (y1 , . . . , ys )E : I/(y1 , . . . , ys )E.
2
52 Local ring theory
Proof In Section I.1 we have proved the implication one way (Corollary I.1.6).
It remains to prove that if hdR R/(x, y) ≤ 2 for every x, y ∈ R, then R is a unique
factorization domain.
First we see from Corollary I.2.10, that R is a domain, since if we pick x = y =
0, we have the finiteness of the homological dimension of R/(x), which guarantees
that x is not a zero divisor. But then, as in the proof of Corollary I.1.6, we see
that the sequence
0 → (x) : y → R2 → R → R/(x, y) → 0
is exact, which tells us that hdR ((x) : y) = 0. Hence this ideal is projective,
indeed free, hence principal, and we are done. 2
II.7 Multiplicity
In Section II.3, we proved that if R is a local ring and q an ideal of defini-
tion of R (i.e., an ideal which contains a power of the maximal ideal), then for
any finitely generated R-module, E, the function χq (E) defined by χq (E; n) =
length(E/qn E) is a polynomial function. In fact, we defined the dimension of
E (Definition II.3.7) to be the common degree of this polynomial for any ideal
60 Local ring theory
of definition, q, and we saw that it was always less than or equal to the dimen-
sion, d, of the ring, R. In this section, we will take a closer look at the leading
coefficient of these polynomials or, more precisely, the coefficient of the term of
degree d.
First, let us expand a bit on what we did in Section II.3 on polynomial
functions. Define the polynomials Xi in Q[X] by
X X(X − 1) · · · (X − i + 1)
= .
i i!
It is easy to show that these polynomials form a linear basis for the polynomial
ring, Q[X], over the rationals, Q. They also help to explain why the term “poly-
nomial” enters into the definition of “polynomial function”
in Section II.3: we
think of the binomial coefficient ni as the polynomial Xi evaluated at n.
If f (n) is an integer-valued polynomial function, then this means that there
are integers a0 , . . . , ad such that
n n
f (n) = a0 + a1 + · · · + ad
1 d
for all n sufficiently
d large
or that, for sufficiently large n, the function, f , and
the polynomial, i=0 ai Xi , agree. Notice that while the term ad is an integer,
the coefficient of X d is ad /d!. We will use this observation in our definition of
multiplicity.
While we will not be making use of many of the convenient properties of the
polynomials, Xi , we point out two of them (without proof) that the reader may
explore if moved by curiosity.
(a) −X i = (−1)i X+i−1
i ;
X+Y i X Y
(b) i = j=0 j i−j .
Employing these, a seemingly complicated identity such as
n+i−j−1 n n+i−1
(−1)j =− ,
j>0
i−j j i
Theorem II.7.2 Let E be a finitely generated module over the local ring, R,
and let q be an ideal of definition generated by x1 , . . . , xs . Then
Before we outline a proof of this result, we point out that the number, s, must
be at least equal to d, the dimension of R (Section II.3). Therefore, the result is
most significant when s = d, that is, when x1 , . . . , xs is a system of parameters,
and dim(E) = d.
C : 0 → Cs → Cs−1 → · · · → C1 → C0 ,
where Ci = Λi Rs ⊗R E.
We now consider the graded ring, R, defined by
R = R/q ⊕ q/q2 ⊕ · · · = qn /qn+1 .
n≥0
C : 0 → C s → C s−1 → · · · → C 1 → C 0
(l)
Therefore, for sufficiently large l, we must have all Hi (C) = 0. If we do this for
each i, we arrive at an integer l0 with the property that for all l ≥ l0 , we have
(l)
Hi (C) = 0 for all i. It is this fact that we exploit in order to obtain our result.
And we exploit it by writing explicitly what this means, namely, the following
sequence is exact for sufficiently large l:
which in turn implies that Hi (C) = Hi (C/C(l) ) for all sufficiently large l. Thus
these two complexes have equal Euler–Poincaré characteristics. However, it is
clear that the terms of the complexes C/C(l) are of finite length over R, so that
we have
From the above result follow a number of others of interest. We will list some
of them, and refer the reader to Reference [9], sections 4–6, and Reference [78]
for proofs and more detail.
This next result is an immediate corollary of the above theorem, and is the
aim of the foregoing discussion.
χ(H(E(x1 , . . . , xd )) = eE (q).
R-module, then
d−1
eE (q) = length(E/qE)−length(qd−1 E : xd /qd−1 E) − e(ql−1 E:xl /ql−1 E) (q/ql )
l=1
where ql = (x1 , . . . , xl ), l = 0, . . . , d − 1.
This theorem shows us that the multiplicity, eE (q), always exceeds the length
of E/qE. Further investigation into what accounts for this difference led to the
study of Macaulay modules and rings, a notion that we now define.
Definition II.7.5 An R-module, E, is called a Cohen–Macaulay module
(C–M for short) if codimE = dimR. The ring, R, is Cohen–Macaulay (C–M
for short) if, when considered as a module over itself, it is a C–M module.
The following two results provide us with some insight into the relationship
between the property of being C–M, unmixedness, and systems of parameters.
These are theorem 5.6 and proposition 5.7 of Reference [9].
Theorem II.7.6 If E is a C–M module, and x1 , . . . , xs s ≥ 0 is a set of
elements in m such that height(E/(x1 , . . . , xs )E) = s, then the submodule (0)
in E/(x1 , . . . , xs )E is unmixed (that is, every prime ideal belonging to (0) in
E/(x1 , . . . , xs )E is of height s) and x1 , . . . , xs is an E-sequence if s ≥ 1.
If E is an R-module having the properties that dimE = dimR and for every set
of elements x1 , . . . , xs in m such that height(E/(x1 , . . . , xs )E) = s, the submodule
(0) in E/(x1 , . . . , xs )E is unmixed, then E is a C–M R-module.
Theorem II.7.7 Let E be an R-module (dimR > 0). Then the following
statements are equivalent:
(a) E is a C–M module.
(b) If x1 , . . . , xd is a system of parameters for R, then Hi (E(x1 , . . . , xd )) = 0
for all i > 0.
(c) If x1 , . . . , xd is a system of parameters for R, then x1 , . . . , xd is an E-
sequence.
(d) There exists a system of parameters for R which is an E-sequence.
Finally, we have the fact that the equality of length and multiplicity
characterizes the property of being C–M ([9], theorem 5.10).
Theorem II.7.8 Let R be a local ring (dimR > 0) with maximal ideal m, let
E be an R-module, and let x1 , . . . , xd be a system of parameters for R. Then the
following statements are equivalent:
(a) E is a C–M module.
(b) The system of parameters x1 , . . . , xd is an E-sequence.
(c) If (y1 , . . . , yd ) = (x1 , . . . , xd ), then codimE/(y1 , . . . , yd−1 )E > 0.
(d) If q = (x1 , . . . , xd ), then eE (q) = length(E/qE).
64 Local ring theory
d
χ(R/p, R/q) = (−1)i length(TorR
i (R/p, R/q));
i=0
d
χ(M, N ) = (−1)i length(TorR
i (M, N )).
i=0
Without going into detail, suffice it to say that his choice of definition was
informed by the desire to be able to prove a “Bézout Theorem.”
The classical conjectures that immediately emerged are:
1. dim(M ) + dim(N ) ≤ dim(R).
2. (Non-negativity) χ(M, N ) ≥ 0.
3. χ(M, N ) > 0 if and only if dim(M ) + dim(N ) = dim(R).
Equivalently, one can write:
1. dim(M ) + dim(N ) ≤ dim(R).
2. (Vanishing) If dim(M ) + dim(N ) < dim(R), χ(M, N ) = 0.
3. (Positivity) If dim(M ) + dim(N ) = dim(R), χ(M, N ) > 0.
In 1985, P. Roberts and H. Gillet-C. Soulé, using K-theory methods, local
Chern classes [74, 75] and Adams operations on Grothendieck groups of com-
plexes [44], succeeded in proving Vanishing. (J.-P. Serre had already proven
that dim(M )+dim(N ) ≤ dim(R).) In 1997, O. Gabber proved it is always the
case that χ(M, N ) ≥ 0. (The best reference for this is the expository paper by
M. Hochster [52], and may be obtained on his web page. His references are to ref-
erences [11] and [36].) Therefore, what remains is to prove the Positivity. Almost
every method of attack on this problem is tied to the following fundamental fact.
The homological conjectures 65
Then
χ(K ⊗ M ) = eM (I),
where eM (I) is the Samuel multiplicity, that is, a slight generalization of the
multiplicity defined in the last section (Section II.7), and discussed in detail in
References [9] and [78].
Usually the proof requires the use of a spectral sequence argument, or else the
type of argument we used in the last section. However, a particular case that is
very suggestive is the following:
Consider, as usual, R, p and q, with length(R/p ⊗ R/q) < ∞. Suppose that q
is generated by a regular sequence x1 , . . . , xs . Then
dim(R/q) = dim(R) − s,
and therefore
and
Since the Samuel multiplicity is always non-negative, and positive if and only if
dim(R/p) = s, the conjectures are true in this special case.
In the equicharacteristic case, that is, in the case that both the local ring, R,
and its residue field, k, have the same characteristic, J.-P. Serre proved the mul-
tiplicity conjecture by means of a “reduction to the diagonal” method, one that
pretty much follows the classical discussions of intersection multiplicity in algeb-
raic geometry. If R is a formal power series ring over a field, K[[X1 , . . . , Xd ]], and
M and N R-modules such that M ⊗R N is of finite length, he introduces a new set
of indeterminates, Y1 , . . . , Yd , and considers N as a module over K[[Y1 , . . . , Yd ]].
He then defines a “complete” tensor product, M ⊗ K N over K, as a module over
66 Local ring theory
BR → BS
↑ ↑
R → S
is commutative. Current status: true for rings that contain a field; open
in mixed characteristic.
13. Serre multiplicity conjectures. Suppose R regular of dimension d, and
M ⊗R N of finite length. Then we have
χ(M, N ) = (−1)i length(TorR
i (M, N )) = 0
68 Local ring theory
if dimM +dimN < d, and positive if the sum equals d. Current status:
true for rings that contain a field; open in mixed characteristic (question of
positivity open: non-negativity is always true).
14. The small C–M modules conjecture. If R is complete, then there exists
a finitely generated R-module, M = 0, such that some (or, equivalently,
every) system of parameters for R is a regular sequence for M . Current
status: true in dimensions ≤ 2; other cases, nothing known, not even if the
ring contains a field, and no matter what the characteristic.
The following is a diagram of implications:
13
14 1
8 → 7, 6, 5 → 4 →3
12 → 11, 10 2
9
III
GENERALIZED KOSZUL COMPLEXES
In the preceding chapter, we saw the fundamental role that the Koszul complex
played in the study of certain aspects of local ring theory. One of the main ideas
tied up with that complex is that of regular sequence, with the Cohen–Macaulay
Theorem one of the major results obtained through its use. There is another
theorem, known as the Generalized Cohen–Macaulay Theorem [35], which deals
with the heights of ideals generated by the minors of an m × n matrix. (More
precisely, the classical form of this theorem states that if R is a regular local ring,
then the ideal generated by the p × p minors of an m × n matrix with entries in
R, has height at most equal to (m − p + 1)(n − p + 1).) We can regard the “old”
Cohen–Macaulay Theorem, then, as having to do with the 1 × 1 minors of an
m × 1 matrix, and the Koszul complex as a complex that “approximates” the
resolution of a cyclic module (that is, a module generated by a single element).
With this point of view, it is natural to ask if one can associate a complex to an
arbitrary m×n matrix (m ≥ n) which generalizes the Koszul complex, and which
may be used to prove the generalized theorem mentioned above. This chapter
will focus on this question; in fact, we will develop a whole class of “generalized
Koszul complexes” that will also play a role in Chapter IV.
In our last section of this chapter we take a look at the connection between
leading coefficients of certain Hilbert–Samuel polynomials and the Euler–
Poincaré characteristics of our generalized Koszul complexes. It was natural
at the time that the papers [18, 27] were being written, to try to generalize
the notions of Hilbert–Samuel polynomials and multiplicity to the situation of
finitely generated modules rather than just cyclic modules, that is, to mod-
ules of the form coker(f : Rm → Rn ), m ≥ n, rather than those of the form
coker(f : Rm → R). In the last sections of Reference [27], these generalizations
were introduced, but the new multiplicity was not pursued further. However, in
the mid-1990s, a flurry of work on these notions began (see [43, 58–61]), which
includes more geometric approaches to the subject. (In the more recent literat-
ure, this multiplicity is often referred to as “Buchsbaum-Rim multiplicity.”) As
in Chapter II, we have to say that a detailed study of this topic would lead us
much too far beyond the scope of this book, so we merely indicate what the
original definition was and its relation to the generalized Koszul complex.
section will devote itself almost entirely to these considerations. The free modules
we deal with are always assumed to be nontrivial and finitely generated.
which are exact for all choices of q > 0. When q = 0, the complex is nothing
other than
0→R→0
and its homology in dimension zero is R itself. Since R = S/J, this is precisely the
statement that the original Koszul complex is acyclic, and its zero-dimensional
homology is S/J.
Definition III.1.1 For each free R-module, and each integer q ≥ 0, we define
Λq (F ) to be the complex
• An important observation to make here (it will be used later in this chapter)
is that for q ≥ n = rankF , the length—or dimension—of this complex is equal
to n, and the last term on the left is Sq−n F ⊗ Λn F.
A few standard complexes 71
0 → Λq G → S1 (F ) ⊗ Λq−1 G → · · · → Sq−l (F ) ⊗ Λl G
→ · · · → Sq−1 (F ) ⊗ Λ1 G → Sq (F ) → 0.
The boundary maps in the complexes Λq (ϕ) are the evident analogs of those
in Λq (F ) :
1⊗∆ 1⊗ϕ⊗1
Sp (F ) ⊗ Λk G → Sp (F ) ⊗ G ⊗ Λk−1 G →
m⊗1
Sp (F ) ⊗ F ⊗ Λk−1 G → Sp+1 (F ) ⊗ Λk−1 G,
Of course, while the complexes Λq (F ) are exact for all q > 0, that is not
necessarily the case for arbitrary Λq (ϕ).
As we saw in Chapter I, the graded dual of the symmetric algebra is the
divided power algebra. Thus, if we were to take the linear dual of the complex
Λq (F ) above, we would obtain, for each q > 0, the exact sequence (exact because
Λq (F ) is split exact) Dq (F ) :
0 → R → 0.
they are projective is clear, since the complexes Λq (F ) and Dq (F ) are split exact;
hence the cycles are direct summands of free modules and therefore projective.
To see that they are indeed free, let us observe that if R is any commutative
ring, and F a free R-module, then F = R ⊗Z F0 where F0 is a free Z-module (of
the same rank as F ). Since Λk F = R ⊗Z Λk F0 , Dp (F ) = R ⊗Z Dp (F0 ), Sp (F ) =
R ⊗Z Sp (F0 ), and since the complexes (over any ring) are split exact, we see
first of all that the cycles of Λq (F0 ) and Dq (F0 ) are free (because projective
abelian groups, being summands of free abelian groups, are torsion-free, hence
free by Proposition I.2.28), and then that the cycles of Λq (F ) and Dq (F ) are free,
because they are just the tensor product with R of their integral counterparts.
This property of a functor, that is, that is R-free and can be obtained by tensoring
its integral counterpart by the ring R, is called universal freeness. With the
freeness of these cycles established, how do we go about finding a “convenient”
basis for them?
To this end, we will use a standard procedure in homology: we will construct
a splitting homotopy for the complex Dq (F ), q > 0; this will pick out the basis
for the cycles that we are looking for. One proceeds almost identically for the
complexes Λq (F ); we will omit the discussion of the homotopy in that case, but
simply give a description of the resulting basis.
To construct the desired homotopy, we have to define maps
sqp : Dp (F ) ⊗R Λq−p F → Dp+1 (F ) ⊗R Λq−p−1 F for − 1 ≤ p ≤ q
such that
q
(i) ∂p+1 sqp + sqp−1 ∂pq = 1 for 0 ≤ p ≤ q;
(ii) sqp+1 sqp = 0 for 0 ≤ p ≤ q − 1.
Of course, the maps sq−1
and sqq
are the zero maps.
Since all of the modules we are considering are free, it suffices to define these
maps on the basis elements. For p > 0, we will use the tableau description of
basis elements that we discussed above, and define
xu1 xu2 · · · xup
xj1
sqp xj2
..
.
xjq−p
0 if u1 < j1 ;
xj1 xu1 ··· xup
xj2
=
xj3
if u1 ≥ j1 .
..
.
xjq−p
74 Generalized Koszul complexes
For p = 0, we define
xj1
xj2
sq0 (xj1 ∧ xj2 ∧ · · · ∧ xjq ) = .. .
.
xjq
Now we have to check the conditions (i) and (ii) above. First of all, it is clear
that
xj1
xj2
∂1q sq0 (xj1 ∧ xj2 ∧ · · · ∧ xjq ) = ∂1q . = xj1 ∧ xj2 ∧ · · · ∧ xjq ,
..
xjq
so (i) is verified for p = 0. For p > 0, we have
xu1 xu2 · · · xup
xj1
q q xj2
∂p+1 sp
..
.
xjq−p
0 if u1 < j1
xj1 xu1 · · · xup
xj2
=
q xj3
∂ p+1 if u1 ≥ j1 ,
..
.
xjq−p
while
xu1 xu2 ··· xup
xj1 m
xj2
sqp−1 ∂pq
q
= sp−1 Th ,
..
. h=1
xjq−p
where h runs, in order, through the m distinct elements of {u1, . . . , up } and
xu1 xu2 · · · x h · · · xup
xh
xj1
Th = ,
..
.
xjq−p
h indicates that the element xh is to be omitted.
where x
A few standard complexes 75
Now let us consider the two cases: u1 < j1 and u1 ≥ j1 . In the first case, we
must show that
xu1 xu2 · · · xup
m xj1
sqp−1 Th = xj2 .
..
h=1 .
xjq−p
Since u1 ≤ · · · ≤ up , we see that except when h = u1 , the tableau Th always has
the property that u1 is less than all the subscripts in the first column, so that
sqp−1 (Th ) = 0. Thus, we see that
xu1 xu2 · · · xup
xj1
q q xj2
sp−1 ∂p
..
.
xjq−p
u1
x xu2 ··· xup
xu1
xj1
= sqp−1 .
..
.
xjq−p
xu1 xu2 ··· xup
xj1
xj2
Since this latter term is equal to , we are done.
..
.
xjq−p
In the second case, we must calculate the sum
xj1 xu1 · · · xup
xj2
q xj3
∂p+1
..
.
xjq−p
xu1 xu2 ··· h
x ··· xup
xh
m
xj1
+ sqp−1 .
..
h=1 .
xjq−p
76 Generalized Koszul complexes
Since we are now assuming that u1 ≥ j1 , we see that in each tableau Th for
which h = u1 , h > j1 so that
xu1 xu2 ··· h
x ··· xup
xh
xj1
sqp−1
..
.
xjq−p
xj1 xu1 ··· h
x ··· xup
xh
xj2
= −
..
.
xjq−p
u1
x xu2 ··· xup
xu1
xj1
sqp−1 =0
..
.
xjq−p
u1
x xu2 ··· xup
xu1
xj1
sqp−1
..
.
xjq−p
xj1 u1
x xu2 ··· xup
xu1
xj2
= − .
..
.
xjq−p
A few standard complexes 77
xj1 xu1 ··· xup
xj2
q xj3
As for the other summand, ∂p+1 , we see that
..
.
xjq−p
it is equal to
xj1 xu1 ··· h
x ··· xup
xh
m xj2
h=1 if u1 = j1 ;
..
.
xjq−p
xu1 xu2 ··· xup
xj1
xj2
..
.
xjq−p
xj1 xu1 ··· h
x ··· xup
xh
m
xj2
+ if u1 > j1 .
..
h=1 .
xjq−p
Comparing terms (including signs), we see that we obtain our desired result. The
verification of condition (ii) is straightforward; in fact it is trivial.
With our homotopy in hand, how do we describe the bases we are looking for?
We claim that if we have a basis of Dp (F )⊗Λq−p F, and consider the subset {Tβ }
q
consisting of those elements for which sqp (Tβ ) = 0, then {∂p+1 sqp (Tβ )} generate
the p-cycles of Dq (F ). For if z ∈ Dp (F ) ⊗ Λ F is a p-cycle, write
q−p
z= cα Tα + d β Tβ
α β
where the Tα are the basis elements that map to zero, while the Tβ are those
already described. Since
q
z = ∂p+1 sqp (z) + sqp−1 ∂pq (z),
and z is a cycle, we have
q
z = ∂p+1 q
sqp (z) = ∂p+1 sqp dβ Tβ = q
dβ ∂p+1 sqp (Tβ ).
β β
78 Generalized Koszul complexes
q
Therefore, if the elements {∂p+1 sqp (Tβ )} are linearly independent, they form a
basis for the p-cycles.
That these elements are linearly independent is easy to see in the following
way. If we have a linear combination
q
dβ ∂p+1 sqp (Tβ ) = 0,
β
q
this means that the term z = β dβ sqp (Tβ ) ∈ ker(∂p+1 ). That is, z is a (p + 1)-
cycle, and as such we know z = ∂p+2 sp+1 (z). But since sqp+1 sqp = 0, this tells
q q
us that z = 0. However, we know that each tableau sqp (Tβ ) is a basis element
of Dp+1 (F ) ⊗ Λq−p−1 F, so that if z = 0, we must have that each dβ = 0. This
q
proves the linear independence we are after. (It also shows us that the map ∂p+1
restricted to the image of sp is one-to-one.) Now the set of tableaux {sp (Tβ )} is
q q
precisely the subset of the basis tableaux of Dp+1 (F ) ⊗ Λq−p−1 F of the form
xu1 xu2 · · · xup+1
xj1
xj2
T(u,j) =
..
.
xjq−p−1
with u1 ≤ u2 ≤ · · · ≤ up+1 and u1 < j1 < j2 < · · · < jq−p−1 . Since it is clear
that every element sqp (Tβ ) is of the form just described, it suffices to show that
every tableau T(u,j) of that form is sqp (Tβ ) for some tableau Tβ . But we see that
xu2 xu3 · · · xup+1
xu1
T(u,j) = sqp xj1 , and so we are done.
..
.
xjq−p−1
q
Now, though, we are after a good way to describe the elements ∂p+1 (T(u,j) )
xu1 xu2 · · · xup+1
xj1
xj2
for such tableaux T(u,j) = . The convention
..
.
xjq−p−1
generally adopted
is to again use tableau notation,
and to denote this boundary
xu1 xu2 · · · xup+1
xj1
xj2
by the tableau , where we see that the rows are
..
.
xjq−p−1
weakly increasing, and the column is strictly increasing. Such a tableau has
A few standard complexes 79
a special name; it is called a standard tableau. Notice, then, that this tableau
stands for the sum of a number of terms in Dp (F ) ⊗ Λq−p F obtained essentially
by diagonalizing the terms in the first row, and multiplying them with the term
in ΛF represented by the product of the terms in the first
column starting with
x1 x1 x2 x2
xj1 . Thus, for example, the tableau x3 represents the element
x4
x1 x2 x2 x1 x1 x2
x1 x2
+ in D3 (F ) ⊗ Λ3 F.
x3 x3
x4 x4
The corresponding discussion for the complexes Λq (F ) considers basis ele-
ments of Sq−l ⊗ Λl F which are tableaux of the form
xu1 xu2 · · · xul
xj1
xj2
,
..
.
xjq−l
where now we have u1 < u2 < · · · < ul and j1 ≤ j2 ≤ · · · ≤ jq−l , and no relation
assumed between u1 and j1 . This means that we are regarding the elements of
the first row as xu1 ∧ xu2 ∧ · · · ∧ xul ∈ Λl F, and xj1 · · · xjq−l as an element of
Sq−l (F ). One then shows that the l-cycles of Λq (F ) can be parametrized by the
set of tableaux
xu1 xu2 · · · xul+1
xj1
xj2
..
.
xjq−l−1
with u1 < u2 < · · · < ul+1 , j1 ≤ j2 ≤ · · · ≤ jq−l−1 and u1 ≤ j1 . Again, to read
off what this element is, we diagonalize the element represented by the first row,
and multiply the first entry with the element of S(F ) represented by the product
of the terms in the first column starting with xj1 . (This procedure is a bit more
straightforward than for the first case as we have here no repeats in the top row.)
Definition III.1.4 The module of p-cycles of the complex Dq F is denoted by
K(p+1,1q−p−1 ) (F ); the module of p-cycles of the complex Λq (F ) is denoted by
L(p+1,1q−p−1 ) (F ).
Remark III.1.5 It should be noted immediately, and it will be used later in
this chapter, that for any free module, F, we have
K(ν+1,1q−1 ) F ∗ ∼
= (L(q,1ν ) F )∗ .
80 Generalized Koszul complexes
l−1
λ1 ⊗ · · · ⊗ λl ⊗ m → (−1)l−j λ1 ⊗ · · · ⊗ λj λj+1 ⊗ · · · ⊗ λl ⊗ m
j=1
+ λ1 ⊗ · · · ⊗ λl−1 ⊗ λl m.
In the case of a graded algebra, Λ, one usually restricts the degrees of the λj
to be positive and, in that case, the complex is referred to as the normalized
bar complex (it is usually assumed that all the λj are homogeneous and that
the module, M, is graded also). We will always make this convention when there
is a grading. Observe, too, that in the graded case, we get a subcomplex by
stipulating that the degree of the λ1 should always be greater than or equal to
some fixed integer, s. Another useful observation is that we may always take as
our module, M, the algebra Λ itself. (It will always be assumed, unless specified
otherwise, that the tensor product is taken over the ground ring, R.)
If B is an arbitrary module over R, we may consider its dual, B ∗=HomR (B,R),
and the two exterior algebras: ΛB and ΛB ∗ . We have seen, in our original
General setup 81
· · · → ΛB ⊗ ΛB ⊗ ΛA → ΛB ∗ ⊗ ΛA → ΛA.
∗ ∗
We remarked above that if our ring Λ is graded, and our module M is a graded
Λ-module, then we can consider graded strands of the bar complex.
In particular, suppose that we are given two integers p ≥ 0 and s > 0. Then
instead of the complex we have just written, we could consider
··· → Λk1 B ∗ ⊗ · · · ⊗ Λkl B ∗ ⊗ Λp+k1 +···+kl A →
% &# $
k1 ≥s ki >0 l
··· → Λ B ∗ ⊗ Λk2 B ∗ ⊗ Λp+k1 +k2 A →
k1
k1 ≥s k2 >0
Λk B ∗ ⊗ Λp+k A → Λp A.
k≥s
Notice that if we choose s = 1, and sum for all p, we get the full (graded)
complex.
We can also do the same thing using ΛB as our ΛB ∗ -module and, given our
map Λf : ΛA → ΛB induced from f : A → B, it is not hard to see that we get
a commutative diagram
··· → Λk1 B ∗ ⊗ Λk2 B ∗ ⊗ Λp+|k| A → Λk B ∗ ⊗ Λp+k A → Λp A
k1 ≥s,k2 >0 k≥s
↓ ↓ ↓
∗ ∗ ∗
··· → Λ B ⊗Λ B ⊗Λ
k1 k2 p+|k|
B→ Λ B ⊗Λ
k p+k
B→Λ B
p
where |k| denotes the sum of the indices ki and the vertical map is the map Λf
tensored with the identity on the appropriate sum of tensor products of copies
of ΛB ∗ .
82 Generalized Koszul complexes
k≥s
It is worth remarking that the freeness of the modules A and B does not enter
into the above discussion; it is enough to assume that our module B is such that
Λn+1 B = 0 to achieve the same result. It may also be amusing to think of the
fact that the composition k≥s Λk B ∗ ⊗ Λp+k A → Λp A → Λp B is zero as the
generalization of Cramer’s Rule from linear algebra. Although we will continue
to use the letters A and B, what we should really keep in the back of our minds
is that A = Rm , and B = Rn , with m ≥ n. So, for example, if m = n, our
complex is simply
· · · → 0 → 0 → Λp R n → Λ p R n ,
while for m = n + 1, we have
0 → Λn−p+1 Rn∗ ⊗ Λn+1 Rn+1 → Λp Rn+1 → Λp Rn .
Up to this point our complexes look fairly svelte, but as soon as the difference
m − n becomes 2 or larger, we get a great many terms that seem extraneous. We
will soon see how this problem is addressed in general, but for now, let us just
look at the case m = n + 3 to see how “fat” our complexes can get to be.
0 → Λn−p+1 Rn∗ ⊗ Rn∗ ⊗ Rn∗ ⊗ Λn+3 Rn+3
Λn−p+1 Rn∗ ⊗ Λ1 Rn∗ ⊗ Λn+2 Rn+3 ⊕
→ Λn−p+1 Rn∗ ⊗ Λ2 Rn∗ ⊗ Λn+3 Rn+3 ⊕
Λn−p+2 Rn∗ ⊗ Λ1 Rn∗ ⊗ Λn+3 Rn+3
Λn−p+1 Rn∗ ⊗ Λn+1 Rn+3 ⊕
→ Λn−p+2 Rn∗ ⊗ Λn+2 Rn+3 ⊕ → Λp Rn+3 → Λp Rn .
Λn−p+3 Rn∗ ⊗ Λn+3 Rn+3
Despite this obesity, the complex always has dimension m − n + 1, the number
suggested (at least in the case when p = n) by the generalized Cohen–Macaulay
General setup 83
Theorem. But it would be more efficient if we could reduce the fat while keeping
the correct dimension. To this end, let us take a closer look at what is going on
and what it is we want.
Λk+l B ∗ ⊗ Λt+l A
Λk B ∗ ⊗ Λl B ∗ ⊗ Λt+l A Λt−k A ,
Λk B ∗ ⊗ Λ t A
The natural question to ask is whether this slimming down can be continued
all along our fat complexes. That is, can we do what is suggested above, and
define a sleek complex associated to any map f : A → Rn and any integer
p≤n:
m
0 → K(m−n,1n−p ) (Rn∗ ) ⊗ ΛA → · · · → K(l,1n−p ) (Rn∗ ) ⊗ Λn+l A → · · ·
→ K(2,1n−p ) (Rn∗ ) ⊗ Λn+2 A → K(1,1n−p ) (Rn∗ ) ⊗ Λn+1 A → Λp A → Λp Rn ?
We notice that this “complex,” too, has dimension m−n+1, but it is considerably
slimmer than the ones we have considered so far.
We have placed quotation marks around the word complex because as yet we
have not defined the boundary maps and verified that they give us a complex.
To define the boundary map, we will focus on the definition of
where the vertical maps are the boundary maps of the complexes Dn−p+l Rn∗
and Dn−p+l−1 Rn∗ , each tensored with the indicated exterior powers of A. The
horizontal maps involve the diagonalization of D(Rn∗ ), and the action of Rn∗ on
ΛA. That is, the upper horizontal map is the composition
Dl (Rn∗ ) ⊗ Λn−p Rn∗ ⊗ Λn+l A → Dl−1 (Rn∗ ) ⊗ Rn∗ ⊗ Λn−p Rn∗ ⊗ Λn+l A
∼
= Dl−1 (Rn∗ ) ⊗ Λn−p Rn∗ ⊗ Rn∗ ⊗ Λn+l A → Dl−1 (Rn∗ ) ⊗ Λn−p Rn∗ ⊗ Λn+l−1 A,
and the lower one is similarly defined. The commutativity of the diagram is easy
to verify.
Given that this diagram is commutative, it follows that the image of the
left-hand vertical map is carried into the image of the right-hand vertical
map. But these images are precisely the modules K(l,1n−p ) (Rn∗ ) ⊗ Λn+l A and
K(l−1,1n−p ) (Rn∗ ) ⊗ Λn+l−1 A that we are after. To check that the composition of
two successive such maps is zero is also very easy.
With this discussion, we are now ready to define the families of complexes that
we want to have at our disposal.
Families of complexes 85
where
C1q = Λq F
and
Ckq = Λn−q+s0 G∗ ⊗ Λs1 G∗ ⊗ · · · ⊗ Λsk−2 G∗ ⊗ Λn+|s| F, k ≥ 2,
si ≥1
|s| = si . The maps (except for Λq f : Λq F → Λq G) are the bar complex maps
associated to the action of the algebra ΛG∗ on ΛF.
As the signs of all our maps are quite crucial, we will make clear just what we
mean by “boundary map” in this context. Namely, if a0 ⊗a1 ⊗· · ·⊗ak−2 ⊗x ∈ Ckq
for k ≥ 3, then
∂(a0 ⊗ a1 ⊗ · · · ⊗ ak−2 ⊗ x)
k−3
= a0 ⊗ a1 ⊗ · · · ⊗ ak−2 (x) + (−1)k−i a0 ⊗ a1 ⊗ · · · ⊗ ai ∧ ai+1 ⊗ · · · ⊗ ak−2 ⊗ x.
i=0
where
Λn−q+s1 G∗ ⊗ Λs2 G∗ ⊗ · · · ⊗ Λsk ⊗ Λn−q+|s|+l F,
(q,l)
Ck = k ≥ 1,
si ≥1
Note that our definition does not preclude the possibility that l be negative,
or that q be equal to n+1. In our immediate use of these complexes, these rather
bizarre possibilities will not show up. In the next section, however, we will make
use of this flexibility when we consider various proofs of “generic acyclicity”
(a term that will be defined in that section) of our complexes.
We now have the following proposition.
(q,q−1)
and γ1,k : Ck → Ckq by
(q,q−1) (q,q−1)
To define γ2,k : Tk → Tkq we use the fact that Tk is the image of
n−q ∗ ∗ α n−q+1 ∗ ∗
Λ G ⊗ Dk G ⊗ Λ n+k−1
F →Λ G ⊗ Dk−1 G ⊗ Λ n+k−1
F, while Tkq is the
β
image of Λn−q G∗ ⊗ Dk−1 G∗ ⊗ Λn+k−1 F → Λn−q+1 G∗ ⊗ Dk−2 G∗ ⊗ Λn+k−1 F. We
then note that we have a map, which we will denote by ∂γ : Dl G∗ → Dl−1 G∗ ,
defined as the composition
∆⊗γ
Dl G∗ = Dl G∗ ⊗ R → Dl−1 G∗ ⊗ G∗ ⊗ G → Dl−1 G∗ ⊗ R = Dl−1 G∗ ,
1⊗ω
∂γ
Λn−q G∗ ⊗ Dk G∗ ⊗ Λn+k−1 F → Λn−q G∗ ⊗ Dk−1 G∗ ⊗ Λn+k−1 F
↓α ↓β
∂γ
Λn−q+1 G∗ ⊗ Dk−1 G∗ ⊗ Λn+k−1 F → Λn−q+1 G∗ ⊗ Dk−2 G∗ ⊗ Λn+k−1 F.
Then the map ∂γ = 1 ⊗ ∂γ ⊗ 1 carries the image of α to the image of β, and this
is the map γ2,k that we are after.
As in the case of the classical Koszul complex, the fact that the mapping cones
of these maps are the appropriate complexes is not difficult to prove; again one
simply uses the identification of Λl (F ⊕ R) with Λl F ⊕ Λl−1 F. 2
88 Generalized Koszul complexes
σ 0 : Λq G → Λ q F
by setting
σ0 (y) = y(ξ)(λ).
In order to show that ∂σ0 (y) = µy, we must prove some basic lemmas about the
operation of ΛG∗ on ΛG in general.
measures ΛG.)
Proof We will not give a detailed proof of this, but just an outline. For |α| = 1,
this is easy: it is just the statement (which we used in Chapter II) that the bound-
ary map of the Koszul complex is a derivation. We then proceed by induction
on the degree of α, and assume that α = α1 ∧ α2 where |α1 | = 1. One then uses
the fact that
∆(α1 ∧ α2 ) = ∆(α1 )∆(α2 ) = (α1 ⊗ 1 + 1 ⊗ α1 ) α2i ⊗ α2i
= α1 ∧ α2i ⊗ α2i + (−1)|α2i | α2i ⊗ α1 ∧ α2i
,
Families of complexes 89
Proof Again, we will not prove this; we just indicate that in this case, one can
use induction on the degree of u. 2
Using the formula above, one sees immediately that ∂σ0 (y) = µy, so our
homotopy is under way. That is, we have
∂σ0 (y) = (−1)|yi |(n−1) yi ∧ ξ(yi ∧ λ) = y ∧ ξ(λ) = µy,
because most of the summands disappear in the above sum (since the λ is now to
be considered as sitting inside Λn G, so that multiplication with yi is zero unless
degree of yi is zero).
To proceed with the definitionof the homotopy, we refine our notation a bit.
We have been writing ∆(x) = j xj ⊗ xj to indicate the total diagonal of x.
However, we occasionally want to specify the degrees
of the terms that occur in
the sum, so to do that we will write ∆(x) = j l xjl ⊗ xjr−l to indicate the
degree l of the term xjl , and the degree r − l of the term xjr−l if the element x
is of degree r.
We now define
σ1 (x) = xjl (ξ) ⊗ xjq−l ∧ λ
j l<q
where x ∈ Λq F.
In order to see that this works, we prove the following two lemmas:
q−l
l+t
xjl (ξ)(xjq−l ∧ λ) = (−1)t xjq−l−t ∧ xjl+t (ξ)(λ).
j j t=0
l
90 Generalized Koszul complexes
and the conclusion follows (because we know what happens to the alternating
sum of binomial coefficients). 2
Now it is easy to see that σ1 gives us what we need.
The next step is the “generic” one; that is, once we get this one, the others all
are of the same type. We define
σ2 (β ⊗ x) = β ⊗ xjn+s−q+l (ξ) ⊗ xjq−l ∧ λ
j l<q−s
for β ∈ Λn−q+s G∗ , x ∈ Λn+s F. There are two “tricks” to showing that this works:
one is to recognize that
β(xjn+s−q+l )(ξ) = β ∧ xjn+s−q+l (ξ)
because ξ is of degree n (we are applying the
corollary to the measuring formula).
Then when we compute the boundary of l<q−s β ⊗ xjn+s−q+l (ξ) ⊗ xjq−l ∧ λ,
Families of complexes 91
we can allow l = q − s, since the zero degree term in the middle cancels out in
the boundary. But then, since the degree of x is greater than n (since s ≥ 1),
the term x(ξ) = 0, so when we apply our lemma, we see that this definition of
σ2 works.
Now it is easy to see that for k ≥ 2, we may define
σk+1 (β0 ⊗ · · · ⊗ βk−1 ⊗ x) = β0 ⊗ · · · ⊗ βk−1 ⊗ xjn+|s|−q+l (ξ)
j l<q−|s|
⊗ xjq−l ∧λ
for β0 ∈ Λn−q+s0 G∗ , βi ∈ Λsi G∗ , for i ≥ 1, and x ∈ Λn+|s| F (|s| = i≥0 si ).
We therefore state (without proof) the following theorem.
σ0 (y) = y(ξ)(λ);
σ1 (x) = xjl (ξ) ⊗ xjq−l ∧ λ;
j l<q
and
σk+1 (β0 ⊗ · · · ⊗ βk−1 ⊗ x) = β0 ⊗ · · · ⊗ βk−1 ⊗ xjn+|s|−q+l (ξ)
j l<q−|s|
⊗ xjq−l ∧ λ
where the right arrow is the k-fold diagonalization of Dk G∗ . (We observe that
q
the latter term is a summand of Ck+2 .)
It is a relatively straightforward calculation to see that αq , thus defined, is a
map of complexes.
To define the maps θq : C(q; f ) → T(q; f ) requires a bit of extra preparation.
In dimensions 0 and 1 it is of course defined to be the identity and denoted by
q q
θ−2 and θ−1 and as in the definition of the map αq , we will denote by θkq the
q q
map that sends Ck+2 to Tk+2 . But now we introduce some notation to facilitate
the definition.
Notation
Assume that a fixed basis of G∗ is given, say y1 , . . . , yn . An element yj1 ∧ · · · ∧ yjl
will be written either as j1 ∧ · · · ∧ jl or j1 · · · jl or simply as J. In short, the index
on a basis element will be used to denote that element (as is the practice when
working with tableaux), and products of elements will be denoted by products of
their indices. For l = 1, we will usually write j or j1 instead of J. When working
with products in the divided power algebra, we will use tableau notation in order
to avoid confusion about whether juxtaposition means the usual product within
that algebra, or the divided power when there are repeats. For example, we will
(2)
write, for yu1 yu2 , with u1 < u2 , the tableau [u1 u1 u2 ].
We will use freely the standard basis, consisting of standard tableaux, for
“hook Weyl modules” (as described in Subsection III.1.2), and a typical basis
q
element of Tk+2 would be denoted by
j1 u1 · · · uk
j2
.. ⊗x
.
jn−q+1
j1 u1 ··· uk
j2
where x is an element of Λn+k+1 F, and .. is a standard
.
jn−q+1
tableau which stands for the image of the element j2 · · · jn−q+1 ⊗ [j1 u1 · · ·
uk ] ∈ Λn−q G∗ ⊗Dk+1 G∗ in Λn−q+1 G∗ ⊗Dk G∗ under the map which diagonalizes
Dk+1 G∗ → G∗ ⊗ Dk G∗ , and then multiplies Λn−q G∗ into Λn−q+1 G∗ by using
the G∗ factor.
Families of complexes 93
Recall that “standard tableau” means that the indices are strictly increasing
in the column, and weakly increasing in the row. For reasons that will become
apparent later, we will make one more (unusual) convention, which we will use
in this section only, about our use of tableau notation in the case of rows: we will
assume that the tableau is zero if the top row is not weakly increasing as written.
(2)
Thus, in the case of yu1 yu2 with u1 > u2 , we would have to write [u2 u1 u1 ]
to represent it as a tableau.
We are now in a position to define our maps θkq for all k ≥ −2.
Definition III.3.11 (Definition of the map θq ) For k = −2 and −1, we
have already said that the map is to be the identity. For k = 0, and Y = J ⊗ x,
with J a basis element of Λn−q+s G∗ and x ∈ Λn+1 F, we define
θ0q (Y ) = (−1)(s−1)(n−q+1) js · · · jn−q+s ⊗ j1 · · · js−1 (x).
For q > 0, and Y = J ⊗ u1 ⊗ · · · ⊗ uk ⊗ x with all the ui basis elements of degree
one, and J still of degree n − q + s, we define
js uk · · · u1
js+1
θk (Y ) = (−1)(s−1)(n−q+1)+(s−1)k . ⊗ j1 · · · js−1 (x).
..
jn−q+s
It is essential to remember here that the tableau is to be read as equal to zero if
the row is not standard. Assume that the map θlq has been defined on elements
Y = J ⊗U1 ⊗· · ·⊗Ul ⊗x for Ui basis elements of arbitrary degree, and l < k, and
that θkq has been defined on elements Z = J ⊗U1 ⊗· · ·⊗Uk ⊗x with U1 , . . . , Ut basis
elements of arbitrary degree (we make the convention that U0 = J), Ut+2 , . . . , Uk
basis elements of degree 1, and Ut+1 basis element of degree st+1 ≤ r (r > 0).
We now let Z = J ⊗ U1 ⊗ · · · ⊗ Ut+1 ⊗ · · · ⊗ Uk ⊗ x with the basis element
Ut+1 = v ∧ W, v of degree 1, all of the basis elements Ut+2 , . . . , Uk are of degree
1, and degree(W ) = r. Define
θkq (Z ) = θkq (B + (−1)(k−t−1)r E)
where B = J ⊗ U1 ⊗ · · · ⊗ Ut v ⊗ W ⊗ Ut+2 ⊗ · · · ⊗ Uk ⊗ x, and E = J ⊗ U1 ⊗
· · · ⊗ Ut ⊗ v ⊗ Ut+2 ⊗ · · · ⊗ Uk ⊗ W (x).
Notice that in position t + 1, the elements B and E are of degree less than
or equal to r, while the terms of higher index are not affected in degree. From
this we see immediately (using a simple induction proof) that θkq (Z ) = 0 unless
Ut+2 ≥ · · · ≥ Uk .
While it is trivial to show that the map αq is a map of complexes, it is not
trivial to prove that the same is true of θq . It is at this point that we seriously
invoke the suggestion that the reader look at Reference [12] for details. In any
event, it is quite straightforward to show that the composition θq αq is equal to
the identity on the complex T(q; f ).
94 Generalized Koszul complexes
This fact is actually enough for us to transfer the homothety homotopy from
the fat to the slim complexes. In fact, we have the following general lemma that
will give us the result we want.
∂X τi + τi−1 ∂X = µ.
∂X τi = ∂X θi+1 si αi = θi ∂Y si αi = θi (µ − si−1 ∂Y ) αi
= µ − θi si−1 ∂Y αi
while
With this lemma, we now know that our slim complexes carry the desired
homotopy.
Although, as we said, we will not give a proof of the following fact, we state
it for the sake of completeness.
and only if Supp R/I ∩ Ass(A) = 0, and this is if and only if T 0 (A) = 0. Hence
we have our result for depth(I; A) = 0.
Let c(A) be the smallest integer r such that T r (A) = 0, and let us assume that
depth(I; A) = d > 0. We want to show, as we have in the case depth(I; A) = 0,
that d = c(A). Let α1 , . . . , αd be a maximal regular A-sequence contained in I,
and consider the exact sequence
0 → A → A → A/α1 A → 0.
We know that depth(I; A/α1 A) = d − 1 so, by induction, we have that
c(A/α1 A) = d − 1. From the exactness of
T i (α1 )
· · · → T i−1 (A/α1 A) → T i (A) → T i (A) → T i (A/α1 A) → · · ·
it follows that T i (α1 ) is a monomorphism for all i ≤ c(A/α1 A). But if T i (α1 ) is
a monomorphism, then so is T i (α1h ) for all h. However, SuppT i (A) ⊆ Supp R/I,
so that the radical of the annihilator of T i (A) contains the radical of I. Since
α1 ∈ I, some power of α1 must annihilate T i (A), and since we are assuming that
α1h = T i (α1h ), we cannot have that T (α1h ) is a monomorphism unless T i (A) = 0.
From this we see that T i (A) = 0 for all i ≤ d − 1. To see that T d (A) = 0, we
simply observe that
0 → T d−1 (A/α1 A) → T d (A),
is exact, and that tells us what we want to know, that is, that c(A) = d. This
concludes the proof of the first statement of our theorem.
Before we prove the second part, notice that the functors ExtiR (R/I, ∗) and
ExtiR (R/radI, ∗) each satisfy the conditions of our theorem with respect to
either of the ideals, I and radI. Thus we see immediately that depth(I; A) =
depth(radI; A).
Now for the proof of the second statement: we assume that there is a fixed
R-module, M, such that T 0 (∗) = HomR (M, ∗). If we take any prime ideal p, and
use the fact that
Supp HomR (M, R/p) = Supp T 0 (R/p) ⊆ Supp R/I,
we see that Mp = 0 for all p not in Supp R/I, that is, Supp M ⊆ Supp R/I.
Therefore, both {T i (∗)} and {ExtiR (M, ∗)} are homology functors satisfying the
conditions of the theorem. Suppose that d = depth(I; A). If we let J = Ann(M )∩
Ann(T d (A)) ∩ I, and apply the fact that Supp R/I contains both Supp M and
Supp T d (A), we see that radJ = radI. Therefore depth(J; A) = depth(I; A) = d,
and we can find a maximal A-regular sequence α1 , . . . , αd in J. Now, by the
property of being homology functors, and from the fact that the αi annihilate
T d (A) and ExtdR (M, A), we have
T d (A) ∼
= T 0 (A/(α1 , . . . , αd )A) = HomR (M, A/(α1 , . . . , αd )A) ∼
= ExtdR (M, A),
and the proof of the theorem is complete. 2
Depth-sensitivity of T(q; f ) 97
between the first and second rows are the operation of F on SG through the
map f and, in the third column, we have the induced maps on the indicated
modules. We are using the fact that if M = coker(f ), for any map f : C → D,
then the sequence
Sν−1 D ⊗ C → Sν D → Sν M → 0
is exact for all ν.
Therefore we see that both T i and U i are homology functors satisfying the con-
ditions of our theorem, or we will as soon as we show that SuppN = Supp M =
Supp R/I(f ). Since M = 0 implies that N = 0, it suffices to show that M = 0
implies that N = 0. By localizing at primes, we may assume that R is local,
and then by reducing modulo the maximal ideal, we may assume that R is a
field. In that case, all the modules are finite-dimensional vector spaces, so we
may assume that G has a basis consisting of a basis of M together with a basis
for the image of f. It is easy, then, to construct an element of Λt G ⊗ Sν M which
is not in the kernel of ϕt−1,ν+1 . But, since the kernel of ϕt−1,ν+1 contains the
image of ϕt,ν , this gives us an element not in the image of ϕt,ν , and we are done.
When m = n, it is easy to see, using the same kinds of identifications used
above, that U 0 (A) ∼= HomR (Q, A), where Q = coker(Λn−q G∗ → Λn−q F ∗ ). Since
m = n, it is easy to prove that SuppQ = Supp R/I(f ). When m = n + 1, similar
identifications show that U 0 (A) = HomR (Q , A), where Q = coker(Λm−q F →
Λm−q G). Again, it is easy to see that the support of Q is equal to that of R/I(f ).
With these observations, we are able to apply our theorem to the functors T i
and U i . In fact, we will state a number of results without proof, as the details
of proof may be found in Reference [27]. (Although the proofs in Reference [27]
relate to the complexes C(q; f ), our discussion above is sufficient to see that they
apply equally to the complexes T(q; f ).)
Theorem III.4.3 Given our map f : F → G and an R-module M such that
M/(I(f ))M = 0, we have for each q with 1 ≤ q ≤ n, the following facts:
1. depth(I(f ); M ) = the smallest integer r for which H r (T(q; f ), M ) = 0, and
furthermore,
H d (T(q; f ), M ) = ExtdR (coker(Λq f ), M )
where d = depth(I(f ); M ).
2. m − n + 1 − depth(I(f ); M ) is equal to the largest integer r for which
Hr (T(q; f ), M ) = 0. Furthermore
ExtdR (coker(Λm−q f ), M ) if m = n or n + 1;
Hm−n+1−d (T(q; f ), M ) =
ExtdR (Γn−q+1,m−n−1 , M ) if m > n + 1.
Corollary III.4.4 supM depth(I(f ); M ) ≤ m − n + 1, where M runs through
all R-modules such that M/I(f )M = 0. In particular, depth(I(f ); R) ≤ m−n+1.
Another kind of multiplicity 99
We should note that there is a stronger result that can be proven, namely:
supp dim Rp ≤ m − n + 1, where p runs through all minimal primes containing
I(f ). The reader may find this in Reference [27], theorem 3.5.
Proof The proof of this result has the same flavor as that of Theorem II.7.2.
The important observation that gets the proof rolling is that we have the exact
sequence of S(Rm )-modules
image of τm ), and therefore all of its graded components are zero from a certain
integer on.
The rest of the proof is easy to follow. 2
Definition III.5.2 For any polynomial function, ϕ, set
u(ϕ) = (degϕ)! × (the leading coefficient of ϕ).
We now have a sequence of results from Reference [27], section 3, which enable
us to define our generalized multiplicity and relate it to our Koszul complexes.
The first of these is Reference [27], theorems 3.3 and 3.4.
Theorem III.5.3 Let f : Rm → Rn be a map, and E and R-module such that
length(coker(f ) ⊗R E) < ∞. Then u(Pf (E; ν)) and degPf (E; ν) − n + 1 depend
only on the R-modules E and coker(f ). If R is a local ring, then degPf (E; ν) =
n − 1 + dimE.
We also have the following corollary (corollary 3.6 of Reference [27]).
Corollary III.5.4 Let R be a local ring, and f and E as above. Then m − n +
1 ≥ dimE, and hence m ≥ degPf (E; ν).
Definition III.5.5 Given a module, M , of finite length over a local ring, R,
choose an exact sequence (a presentation) Rm → Rn → M → 0. Then for each
R-module, E, the product (dimR + n − 1)!× (the coefficient of the term of degree
n−1+dimR in the polynomial Pf (E; ν)) is a non-negative integer which depends
only on M and E. We call this non-negative integer the multiplicity of M with
respect to E, and denote it by eE (M ).
In the case when E = R, we write eR (M ). When M = R/q with q an ideal of
definition, we retrieve our older definition of multiplicity (except that we have it
written now as eE (R/q) instead of eE (q)).
Definition III.5.6 In view of Corollary III.5.4, we call a map f : Rm → Rn a
parameter matrix for E if length(coker(f ) ⊗R E) < ∞ and m − n + 1 = dim E.
The following proposition sums up where we are to this point (Reference [27],
proposition 3.8).
Proposition III.5.7 Let f : Rm → Rn be a parameter matrix for E, and let
M = coker(f ). Then
1. eE (M ) = q (−1)m−q length(Hm−q (K(τm , (S(Rn ) ⊗R E)ν+q ))) for ν suffi-
ciently large.
2. eE (M ) ≥ 0 and eE (M ) = 0 if and only if dimE < dimR.
3. If 0 → E → E → E → 0 is an exact sequence, then eE (M ) = eE (M ) +
eE (M ).
In section 4 of Reference [27], there is a large amount of calculation all of which
is directed at proving the following theorem (see Reference [27], theorem 4.2).
102 Generalized Koszul complexes
In the previous chapter, we saw a whole class of complexes associated with a given
map, f , of finite free modules. We proved that all of them were sensitive to the
depth of the ideal generated by the maximal minors of f , and gave a necessary
and sufficient condition for them to be finite free resolutions of the cokernels
of the exterior powers of f in terms of the depth of the ideal generated by the
maximal minors of f .
In this chapter, we deepen our understanding of what it means for a complex
of free modules to be a resolution, that is, we confront the problem: what does
it mean for a complex of free R-modules to be exact? Given that the underlying
notion of homology, whether it be topological or algebraic, is the exact sequence,
it seems reasonable to ask just what this notion means in everyday terms. Of
course, when we say “everyday terms” we mean in terms of the data that we are
given to hand: in the case of free resolutions, the matrices that represent the
boundary maps of free complexes. This, after all, is what we would look at if
we had a complex of vector spaces.
Of course, there is a vast difference between modules which have finite free
resolutions and those with infinite ones, so it seems reasonable to start with
the more familiar, that is, the ones whose resolutions are finite. Another reason
that compels us to study finite resolutions is that the most familiar rings are
regular: think of the power series rings and polynomial rings. (A commutative
noetherian ring is called regular if its global dimension is finite.) And if one is to
exploit the finiteness of a resolution, one natural starting point is to make use of
the “last” (left-most) matrix that occurs in it. So, from a general point of view,
addressing the question of the structure of finite free resolutions suggests that the
final matrix of the resolution be a jumping-off point. In addition to this purely
theoretical consideration, some problems, such as the Lifting Problem (which
will be explained in this chapter), very much indicated that the last matrix of
a finite free resolution was of significance.
We will see that the complexes studied in Chapter III help us to get some
insight into the minors of the matrices that occur in a resolution. At certain
crucial points, these complexes or their duals will play a role. We will also see
that while our methods carry us some distance toward our goal, there is still a
great deal more to be understood. But first we have to get down to the basics
104 Structure theorems for finite free resolutions
Proof We first do the “if” part. Since I(ψ) = R contains a non-zero divisor,
rank(ψ) does not change when localizing. Hence we assume that R is local and
prove that coker(ψ) is free of rank equal to rankF − rank(ψ). If we call r the
rank of ψ, I(ψ) = R implies that some r × r minor of ψ is invertible. Hence we
may choose bases for E and F such that ψ has block matrix
AB
C D
with A equal to the r × r identity matrix, Ir . Elementary row and column oper-
ations show that B and C may be assumed to be zero. But then D must be
zero as well, otherwise the rank of the overall matrix would exceed r. It is now
obvious that the cokernel of
Ir 0
0 0
is free of rank equal to rankF − rank(ψ).
Now we do the “only if” part. Since coker(ψ) is projective, we get F ∼
=
coker(ψ) ⊕ im(ψ), E ∼
0
= ker(ψ) ⊕ im(ψ), and ψ = idim(ψ) ⊕ ker(ψ) → coker(ψ) .
Let m, n, t be the ranks of E, F , coker(ψ), respectively. If t = n, then ψm = 0
for every maximal ideal m, hence ψ = 0 and I(ψ) = R, as expected. If t < n,
then ψm always has the matrix
In−t 0
,
0 0
hence det In−t = 1 ∈ I(ψm ) and I(ψm ) = (I(ψ))m = Rm for every m. Again
R = I(ψ), as required. 2
The above proposition reaffirms that the rank of R/(e) (of Example IV.1.4)
ψ
is not well defined since it has a presentation R → R → R/(e) → 0 such that
106 Structure theorems for finite free resolutions
is split exact after localization at m (split exact meaning exact with the further
property that im(fi+1 ) is a direct summand of Fi for every i), that the rank of
(coker(fk ))m is equal to
(−1)i−k+1 rankFi
i≥k−1
(a quantity independent of m), and that the projective R-module coker(fk ) has
well-defined rank. Therefore Proposition IV.1.5 applies and I(fk ) = R, as wished.
Now we can prove (1): invert all non-zero divisors, getting I(fk ) = R for all
k, and use Proposition IV.1.6.
As for the proof of (2), take a prime ideal p ⊇ I(fk ) such that codim(Rp ) =
depthI(fk ) (a prime associated to a maximal regular sequence of I(fk )). We
claim that codim(Rp ) ≥ k. To see this, localize F at p: (coker(fk ))p is not free
(by Proposition IV.1.5), hence hdR (coker(f1 ))p is at least k (we cannot shorten
Fp ). Again by Theorem II.4.12, it follows that codim(Rp ) ≥ k, as claimed.
This completes the proof of the “only if” part.
Now we prove the “if” part. We may assume that R is local, with maximal ideal
m (for exactness is a local property, and homology commutes with localization).
Let d indicate the depth of m. For every k > d, condition (2) implies I(fk ) = R.
In particular, coker(fd+1 ) is a free module by Proposition IV.1.5, say Fd+1 . If
fd stands for the map Fd → Fd induced by fd , F can be viewed as the result of
glueing together the following complexes:
fd+1
C : 0 → Fn → · · · → Fd+1 → Fd → Fd → 0
and
f fd−1
F : 0 → Fd →
d
Fd−1 → Fd−2 → · · · → F0 ,
both of which have to be proven to be exact.
The exactness of C comes immediately from Proposition IV.1.6.
The exactness of F (which satisfies the same conditions as F, but with length
d equal to the depth of m) comes by induction on dim R.
If dim R = 0, then d = 0, the complex is trivial and we are through. If
dim R > 0, since localization preserves all assumptions, we know by induction
that F turns out to be exact whenever localized at a non-maximal prime ideal
of R. Hence H(F ), the homology of F , is all supported in m, m is an associated
prime of H(F ), and depthH(F ) = 0. The conclusion now results from the
so-called Acyclicity Lemma due to C. Peskine and L. Szpiro, which we quote
below in a slightly weaker form than the original. 2
Proof Cf. Reference [69] (and also [23, lemma 3] and [41, lemma 20.11]).
2
We remark that both Proposition IV.1.6 and Theorem IV.1.7 above lend them-
selves to several generalizations (cf., e.g., [23]). In particular, the following holds
([23], corollary 2).
Proposition IV.1.9 Let M = 0 be an R-module, and let
ϕn ϕ1
A: 0 → Fn → Fn−1 → · · · → F1 → F0
be a complex of finitely generated projective R-modules of well-defined rank. Then
A ⊗R M is exact if and only if for all k = 1, . . . , n
(a) rank(ϕk+1 ⊗R M ) + rank(ϕk ⊗R M ) = rank Fk and
(b) I(ϕk ⊗R M ) contains an M -sequence of length k or I(ϕk ⊗R M ) = R.
We include this last result because it represents a curiosity in relation to an old
question in homological algebra, namely, the “rigidity problem”. Roughly put,
the rigidity question is this: If N and M are two modules of finite homological
dimension, and if TorR R
i (N, M ) = 0 for some given i, is it true that Torj (N, M ) =
0 for all j ≥ i? Positive results have been obtained on this problem when R is
regular local, but it is not true in general. Possibly the most recent results can
be found in Reference [54], where the authors examine the case when the ring
is a hypersurface, that is, R = S/(f ), where S is a regular local ring. In any
event, we mention this here because the above proposition seems to be made for
studying the rigidity problem; the resolution, A, above could be the resolution of
N (truncated at dimension i), in which case the Tor we are looking at is just the
homology of A ⊗R M . However, none of the methods of attack on this problem
has made use of this kind of result.
There is an important corollary of Theorem IV.1.7, which is often considered
as a second, independent, criterion of exactness for F. We have already remarked
that exactness is a local property. Hence F is exact if and only if F ⊗R Rp is
exact for every prime ideal p. The second criterion says that we may limit the
set of prime ideals, p, for which exactness of F ⊗R Rp must be checked.
Theorem IV.1.10 F is exact if and only if F ⊗R Rp is exact for all prime
ideals p such that depth(pRp ) < n.
Proof In order to prove the “if” part, it suffices to show that F satisfies condi-
tions (1) and (2) of Theorem IV.1.7. In fact, if (2) holds, (1) follows automatically,
since the ranks of the maps localize and (1) holds by assumption over every
Rp . Hence we just prove (2), by contradiction. If depthI(fk ) < k for some k
in {1, . . . , n}, then there exists a prime ideal p ⊇ I(fk ) associated to a max-
imal regular sequence contained in I(fk ). Hence rank((fk )p ) = rank(fk ) and
depth(pRp ) = depthI(fk ) < k. It follows that
I((fk )p ) = (I(fk ))p ⊆ pRp
Some criteria for exactness 109
and depthI((fk )p ) < k, contradicting what property (2) says about the exact
complex F ⊗R Rp . 2
We end this section with some properties enjoyed not just by F, but by a larger
family of exact complexes.
Let G denote a (possibly infinite) complex of finite free R-modules that is
exact:
gk g1
· · · → Gk → Gk−1 → · · · → G1 → G0 .
Suppose that depthI(gk ) ≥ 1 for every k ≥ 1. (By Theorem IV.1.7, this condition
is automatically satisfied when F is finite.)
radI(g1 ) = · · · = radI(gk ).
Proof (a) For x ∈ radI(gk ), let S = {xi }. Then S ∩I(gk ) = ∅ and I(gk )S = RS .
Since by assumption I(gk ) contains a non-zero divisor, rank(gk ) = rank((gk )S )
and I((gk )S ) = (I(gk ))S . Hence coker(gk )S is projective, by Proposition IV.1.5.
Since G exact implies GS exact, the projectivity of coker(gk )S implies that of
coker(gk+1 )S . For the exactness of
I(gk ) = R I(g1 ).
hdR (coker(g1 )) ≥ k,
Proof By Proposition II.6.1, it suffices to show that hdR (R/(x, y)) ≤ 2 for all
x and y in R.
By Theorem II.5.3, we know that hdR (R/(x, y)) is finite. Hence R/(x, y) has
a finite free resolution, say:
fn f2 (x,y)
0 → Fn → Fn−1 → · · · → F2 → R2 → R → R/(x, y) → 0.
Now, Theorem IV.2.1 says that the following diagram commutes
(x,y)
R2 −→ R
∗
a2 ↓ ↑ a1
R = R,
hence the ideal ima∗2 can be generated by two elements, say I(a∗2 ) = (x , y ),
and (x, y) = a1 (x , y ). That is, (x, y) is equal to (x , y ) times a non-zero divisor
(every element of a regular local ring is a non-zero divisor). If we prove that
hdR (x , y ) = 1, it will follow hdR (x, y) = 1, hence hdR (R/(x, y)) ≤ 2 as required.
But Theorem IV.2.1 (b) says radI(f2 ) = radI(a2 ), which implies (recall
Theorem IV.1.7) depthI(a2 ) = depthI(f2 ) ≥ 2. So depthI(a∗2 ) ≥ 2. It easily
follows that x , y is a regular sequence, and hdR (x , y ) = 1 (the Koszul complex
is a resolution of R/(x , y ), of length 2). 2
We now deal with some topics which in fact led to the first structure theorem.
The construction seen in the previous example is related to the “lifting prob-
lem” of Grothendieck, which was motivated by the Serre conjectures on the
positivity of intersection multiplicity. This lifting problem is known to be not
solvable in general (there are counterexamples due to M. Hochster [50]).
Statement of the General Lifting Problem: Given a local ring S and a
non-zero divisor, x, let R = S/(x). If M is an R-module, does there exist an
S-module M . such that x is a non-zero divisor on M . and M ./xM.∼ = M?
t . t
If M is free of finite type, say M = R , clearly M = S does the job.
Consider next any finitely generated M with a free resolution G, say:
gk g1
· · · → Gk → Gk−1 → · · · → G1 → G0 .
One can construct a sequence
: ··· → G
G .k →
g k
G . g1 .
k−1 → · · · → G1 → G0
just by choosing matrices for the maps gk and lifting these matrices. Suppose we
is a complex. We then
are lucky in our choices of the lifted matrices, so that G
get an exact sequence of complexes:
→G
0→G → G →0.
x
the radical, and Nakayama’s lemma applies. Hence the higher homology vanishes
and the long exact homology sequence leaves us with the short exact sequence
→
0 → H0 (G)
x → H0 (G) → 0,
H0 (G)
f1 = (y1 , . . . , y
n+1 ).
It automatically follows that f1 ◦ f2 = 0, and we are through. Beyond this, that
is, if hdR (M ) > 2, anything can happen.
One would like to understand the arithmetic reasons why G is not always
a complex. A possible approach is to look for generic pairs (A, B), namely, a
ring A and an exact sequence B of finitely generated free A-modules with the
following property: given any ring A and any exact sequence B of free A -
modules with the same ranks of those of B, there exists a map A → A such that
B = B ⊗A A . M. Hochster [51] has shown that the structure theorems of this
chapter are enough to get a generic pair for resolutions of length 2. But the struc-
ture of resolutions of length 3 is far more complicated (cf., e.g. References [71]
and [85]).
Remark IV.2.6 Theorem IV.2.1 can sometimes provide conditions for the
cokernel of a map to have finite homological dimension. We illustrate this in
the following example.
Let K be a field and S the polynomial ring K[Xij ], 1 ≤ i ≤ n, 1 ≤ j ≤ m,
with m ≥ n. Consider the generic map
(Xij )
g : S m −→ S n .
As shown in Section III.4, depthI(g) = m − n + 1 = hdR coker(g).
Also let Ip+1 (g) denote the ideal generated by the (p + 1)-minors of the matrix
(Xij ), and set R = S/Ip+1 (g). The generic map of rank p is the induced map
(X ij )
f : Rm → Rn ,
114 Structure theorems for finite free resolutions
where X ij denotes the image of Xij in R. (Notice that p must be less than n.)
We are going to prove that coker(f ) cannot have finite homological dimension.
If hdR coker(f ) were finite, then coker(f ) would have a finite free resolution
(by Quillen-Suslin); indeed a resolution by free R-graded modules, because R
inherits from S the structure of a graded ring (Ip+1 (g) is a graded ideal, g is a
homogeneous map) and coker(f ) is a graded R-module. Theorem IV.2.1 would
thus give the commutative diagram:
Λp f
Λp Rm −→ Λp Rn
↓ a∗2 ↑ a1
R = R.
1
where r stands for the constant a∗3 . Thus
0
2
Y X 1 Y
=
−X −Y 0 −X
X Y
would give r = , from which Y = rX, a contradiction.
Y −X
· · · → G∗ ⊗ Λn+1 F → Λn F → Λn G ∼
ψ2 ψ1
= R,
where ψ1 = Λn f and, for every k > 0, ψk+1 is the composite map
is zero.
Some calculations then show that the following diagram is commutative:
∗
Λrn−1 fn−1 βn
∗ ∗
Λrn−1 Fn−2 → Λrn−1 Fn−1 → Fn∗ ⊗ Λrn−1 −1 Fn−1
∗
i1 1 ⊗ i2
Λrn fn ψ∗
0→ R∼
= Λ r n Fn → Λrn Fn−1 →2 Fn∗ ⊗ Λrn +1 Fn−1 ,
where i1 and i2 are the appropriate isomorphisms given by the orientations (recall
Remark IV.2.2).
It follows that there exists a unique map
∗
Λrn−1 Fn−2 → Λ r n Fn ∼
= R,
whose dual (clearly verifying property 2) we define to be an−1 .
Suppose now that we have ak+1 and ak ; we want to define ak−1 (assume k ≥ 2).
As before, we see that the composite map
∗
∗ Λrk−1 fk−1 ∗ βk
Λrk−1 Fk−2 −→ Λrk−1 Fk−1 −→ Fk∗ ⊗ Λrk−1 −1 Fk−1
∗
is zero.
Consider the sequence
a γ
0→R→
k
Λrk Fk−1 → Fk∗ ⊗ Λrk +1 Fk−1 , (IV.3.1)
where γ is multiplication by the trace of fk , that is, the element of Fk∗ ⊗ Fk−1
corresponding to fk via the canonical isomorphism
HomR (Fk , Fk−1 ) ∼
= Fk∗ ⊗ Fk−1 .
Proof of the first structure theorem 117
a γ
0→ R →
k
Λrk Fk−1 → Fk∗ ⊗ Λrk +1 Fk−1 ,
so that it suffices to show that the bottom row is exact, to be able to conclude
as we did in the case of an−1 .
In order to prove that (IV.3.1) is a complex, we first show that the composite
map
Λrk fk γ
Λrk Fk → Λrk Fk−1 → Fk∗ ⊗ Λrk +1 Fk−1 (IV.3.2)
is zero. Let {x1 , . . . , xt } be a basis of Fk and {ξ1 , . . . , ξt } the
basis of dual to Fk∗
{x1 , . . . , xt }. Then the trace element of fk can be written as i ξi ⊗ fk (xi ), and
for every y1 ∧ · · · ∧ yrk ∈ Λrk Fk it turns out that
(γ ◦ Λrk fk )(y1 ∧ · · · ∧ yrk ) = γ(fk (y1 ) ∧ · · · ∧ fk (yrk ))
= ξi ⊗ fk (xi ) ∧ fk (y1 ) ∧ · · · ∧ fk (yrk )
i
= ξi ⊗ (Λrk +1 fk )(xi ∧ y1 ∧ · · · ∧ yrk ),
i
rk +1
which vanishes because Λ fk = 0.
Next we prove that I(fk ) = I(ak )I(ak+1 ) (this is the equality announced and
used in Example IV.2.5). As the inclusion ⊇ is straightforward, we just prove ⊆.
The three formulas
(Λrk fk )(xi1 ∧ · · · ∧ xirk ) = ±(j1 , . . . , jrk | i1 , . . . , irk )wj1 ∧ · · · ∧ wjrk
(where {w1 , . . . , ws } stands for a basis of Fk−1 and the coefficients occurring in
the sum are k × k minors of fk ),
ak+1 (1) = ci1 ···irk xi1 ∧ · · · ∧ xirk
and
ak (1) = dj1 ···jrk wj1 ∧ · · · ∧ wjrk ,
inserted into the diagram (commutative by hypothesis)
Λrk fk
Λrk Fk −→ Λrk Fk−1
a∗k+1 ↓ ↑ ak (IV.3.3)
R = R,
yield the following diagram
Λrk fk
xi1 ∧ · · · ∧ xirk −→ ±(j1 , . . . , jrk | i1 , . . . , irk )wj1 ∧ · · · ∧ wjrk
↓a∗k+1
a
ci1 ···irk →
k
ci1 ···irk dj1 ···jrk wj1 ∧ · · · ∧ wjrk ,
and we are done.
118 Structure theorems for finite free resolutions
and since I(ak+1 ) contains a non-zero divisor and im(γ ◦ ak ) is included in a free
module, one necessarily has im(γ ◦ ak ) = 0, that is, the sequence (IV.3.1) is a
complex.
It remains to show that (IV.3.1) is exact. We resort to Theorem IV.1.10 and
localize at primes of depth at most 1. Since k ≥ 2 and depthI(fk ) ≥ 2, I(fk )
blows up after localization (i.e., it becomes the whole ring) and we may assume
(recall that I(fk ) = I(ak )I(ak+1 )). It is now obvious that ak is a monomorphism.
To show exactness at Λrk Fk−1 , it suffices to prove that
(by Theorem IV.1.7; the depth condition is satisfied because of (IV.3.4)). (IV.3.5)
is equivalent to
To verify this latter equality, it is enough to show that (IV.3.2) is exact and
use Theorem IV.1.7, since rank(Λrk fk ) = 1. But the exactness of (IV.3.2) is
now obtained by noting that I(fk ) = R implies coker(fk ) projective (Pro-
position IV.1.5), so that Fk = ker(fk ) ⊕ im(fk ), Fk−1 = im(fk ) ⊕ coker(fk )
and
0
fk = idim(fk ) ⊕ {ker(fk ) → coker(fk )}.
For the well-known property Λ(A⊕B) = ΛA⊗ΛB indicates that (IV.3.2) is made
up of strands involving ker(fk ), im(fk ) and coker(fk ), strands whose exactness
is easy to ascertain.
This completes the proof of Part (a) of Theorem IV.2.1.
d1···rk = 1 and let F denote the free submodule of Fk−1 generated by w1 , . . . , wrk .
We take the composite map
fk π
Fk → Fk−1 → F,
where π stands for the projection, and look at the complex
fn fk+1 π◦fk
0 → Fn → Fn−1 → · · · → Fk+1 → Fk → F. (IV.3.6)
By construction, rank(π ◦ fk ) = rankF = rk = rank(fk ) and I(π ◦ fk ) = I(fk ) =
I(ak+1 ), so that using Theorem IV.1.7 we obtain that (IV.3.6) is exact.
The exactness of (IV.3.6) shows that coker(π ◦ fk ), which is not equal to 0,
has finite homological dimension. Since I(fk ) = R (so that neither coker(fk ) nor
coker(π ◦ fk ) is projective),
hdR coker(π ◦ fk ) = hdR coker(f1 ) − k + 1;
hence k ≥ 2 implies hdR coker(π ◦ fk ) hdR coker(f1 ).
As coker(π ◦ fk ) has finite homological dimension, Theorem II.4.12 applies and
hdR coker(π ◦ fk ) codimR.
On the other hand, since rank(π ◦ fk ) = rankF , the discussion after Theorem
III.4.2 holds for π ◦ fk and
I(π ◦ fk ) ⊆ ann(coker(π ◦ fk )).
As I(π◦fk ) = I(fk ), this means that ann(coker(π◦fk )) contains an ideal primary
to m, codim(coker(π ◦ fk )) = 0 and (again by Theorem II.4.12)
hdR coker(π ◦ fk ) = codimR,
a contradiction.
This finishes the proof of Theorem IV.2.1.
there is something we can do in this case also, to get further information about
these maps.
Let C = (cij ) be the (r + 2) × r matrix of f3 . We will describe below an
(r + 2) × 3 matrix, D = (dis ), whose entries, along with those of C, we use to
describe the matrices of f2 and f1 . More precisely, given the (r + 2) × (r + 3)
matrix (C|D) = (cij |dis ), the description of matrices for f2 and f1 , up to signs,
and multiplied by non-zero divisors, goes as follows.
Description of the 3 × (r + 2) matrix for f2 : The (s, t)-coordinate is the
(r + 1) × (r + 1) minor of (C|D) obtained by erasing the row indexed by t and
the columns indexed by the two elements of the set of indices{r + 1, r + 2, r + 3}
which are not equal to r + s.
Description of the 1×3 matrix for f1 : The (1, s)-coordinate is the (r + 2)×
(r + 2) minor of (C|D) obtained by erasing the column indexed by r + s.
The question is: How does one get the matrix D?
We have seen in Section III.4 that given f : F → G with rankF = m ≥
n = rankG, one can construct a family of depth-sensitive complexes T(q; f ). Set
q = 1 and f equal to the map f3∗ : (Rr+2 )∗ → (Rr )∗ . Then T(1; f3∗ ) looks like:
ψ3 ψ2 f∗
T(1; f3∗ ) : 0 → Rr ⊗ Λr+2 (Rr+2 )∗ → Λr+1 (Rr+2 )∗ → (Rr+2 )∗ →
3
(Rr )∗ ,
Λrn−1 −1 Fn−2
∗
−→ Λrn−1 −1 Fn−1
∗
→ Fn∗ ⊗ Λrn−1 −2 Fn−1
∗
,
If, and this is a big if, (IV.4.3) were exact, we could make the usual identifications
and get b∗k as the vertical arrow induced by the commutative box of the following
diagram:
fk+1 a ∗ γ
Fk+1 → −→ Λrk+1 +1 Fk → Fk+1 ⊗ Λrk+1 +2 Fk
k+1
Fk
↑
Λrk −1 f ∗
Λrk −1 Fk−1
∗
−→ k Λrk −1 Fk∗ → Fk+1
∗
⊗ Λrk −2 Fk∗ .
122 Structure theorems for finite free resolutions
↓ bk
Λrk −1 fk
Λrk −1 Fk −→ Λrk −1 Fk−1
If we prove that (IV.4.4) is exact, it will follow that (IV.4.3) is exact too (here
is a case in which it is easier to deal with a longer sequence). In order to show
the exactness of (IV.4.4), we can first prove that it is a complex and then apply
Theorem IV.1.7.
That γ ◦ ak+1 = 0 and a k+1 ◦ fk+1 = 0 follows in the same way that im(γ ◦
ak+1 ) = 0 was proved in Subsection IV.3.1. The rank conditions can be checked
by localizing at S = {xi }, where x is a non-zero divisor in I(fk+1 ), and proceeding
as at the end of Subsection IV.3.1. As for the depth conditions, since k ≥ 2,
depthI(ft ) is all right for all t ≥ k; depthI(
ak+1 ) ≥ 2 we know from the properties
of ak+1 ; depthI(γ) ≥ 1 follows from Lemma IV.4.1 below.
Proof Assume for a contradiction that there exists x ∈ / radI(γ) such that
xt ∈ I(fk+1 ) for some t. Localize at S = {xi }. As S ∩ I(fk+1 ) = ∅ = S ∩ I(γ),
I(fk+1 ) blows up and I(γ) does not. Hence a contradiction will follow, if one
proves that I(fk+1 ) = R implies I(γ) = R.
By Proposition IV.1.5, I(fk+1 ) = R implies that coker(fk+1 ) is projective;
thus we have Fk+1 = ker(fk+1 ) ⊕ im(fk+1 ), Fk = im(fk+1 ) ⊕ coker(fk+1 ) and
0
fk+1 = idim(fk+1 ) ⊕ ker(fk+1 ) → coker(fk+1 ) .
From the well known property: Λ(A ⊕ B) = ΛA ⊗ ΛB, it follows that the
right-most map of (IV.4.4) is made up of strands involving ker(fk ), im(fk )
and coker(fk ), strands whose careful study shows coker(γ ∗ ) = (coker(fk+1 ))∗ ⊗
Λrk+1 (im(fk+1 ))∗ . Hence coker(γ ∗ ) is projective and Proposition IV.1.5 yields
I(γ ∗ ) = R. 2
↓ bk (IV.4.5)
rk −1 Λrk −1 fk rk −1
Λ Fk −→ Λ Fk−1 .
Remark IV.4.3 A map bn : Fn∗ → Λrn −1 Fn−1 can be trivially added to the
statement of the second structure theorem. Simply define it as the composite
Λrn −1 fn
F∗ ∼
n = Λrn −1 Fn −→ Λrn −1 Fn−1 .
Remark IV.4.4 Although Theorems IV.2.1 and IV.4.2 give such good inform-
ation about the maximal and submaximal minors of every fk , it is very hard to
guarantee the existence of a sequence F with prescribed maps ak and bk .
We have just said in the previous remark that it is hard to recover F from the
maps ak and bk . As a matter of fact, given a concrete resolution F, it may not
be easy to find its maps bk either. This motivates the following.
Notice that the commutativity of the diagram in part (a) 2. in the statement
of Theorem IV.2.1 amounts to
ak (a∗k+1 (z(ε))) = Λrk fk (z) (IV.4.6)
for all z ∈ Λrk Fk . Here ε ∈ Λrk+1 +rk Fk∗ denotes the orientation of Fk , that is, the
chosen isomorphism in HomR (Λrk+1 +rk Fk , R) ∼ = Λrk+1 +rk Fk∗ (cf. Remark I.3.32),
rk+1 ∗
and z(ε) ∈ Λ Fk is the contraction of z on ε.
Notice that the commutativity of the diagram (IV.4.5) in the statement of
Theorem IV.4.2 amounts to
b∗k (x)(ak+1 (1)(ε)) = Λrk −1 fk∗ (x)
for all x ∈ Λrk −1 Fk−1
∗
.
Notice also that it is possible to define maps c∗k : Λrk −1 Fk−1
∗
→ Fk such that
∗
fk (ck (x)) = x(ak (1)), because
fk−1 (x(ak (1))) = 0 for all x ∈ Λrk −1 Fk−1
∗
, (IV.4.7)
124 Structure theorems for finite free resolutions
so that one wonders whether the relatively easier maps ck can replace the
maps bk .
[Proof of (IV.4.7). It suffices to show y · fk−1 (x(ak (1))) = 0 for some non-zero
divisor y ∈ R. Since we have already seen in the proof of Theorem IV.2.1, Part
(a), that I(a∗k+1 ) contains a non-zero divisor, we may choose y to be a∗k+1 (z(ε))
for some suitable z ∈ Λrk Fk . We then have
y · fk−1 (x(ak (1))) = fk−1 (x(ak (y))) = fk−1 (x(ak (a∗k+1 (z(ε))))),
that is (recall (IV.4.6)),
fk−1 (x(Λrk fk (z))),
which is zero, since x(Λrk fk (z)) ∈ Fk−1 belongs in fact to im(fk ).]
Proposition IV.4.5 For the maps bk of the second structure theorem, the
following are equivalent:
(1) b∗k (x)(ak+1 (1)(ε)) = Λrk −1 fk∗ (x) for all x ∈ Λrk −1 Fk−1
∗
to understanding some basic questions, we see that they still leave a great many
questions open. Even where the answers to some of our questions are negative,
the counterexamples do not yield the insight we need in order to fully understand
the phenomena that we want to study.
To point to an even more basic, and hence irksome, question, consider the fact
that a module, M , is completely determined by its presentation (which, for the
sake of retaining our sanity, we will assume to be finite): F1 → F0 → M → 0.
Presumably, then, it should be possible to read off all the information we might
want to have about M from the presentation matrix. For example, what property
does the matrix have to have in order that hdR M be finite? Notice that all the
(limited) information we have obtained from the structure theorems comes from
the assumption that the module is known to have finite homological dimension;
using that, we have managed to drag information about the matrix at the tail
of the resolution to the head (i.e., the presentation matrix). We know that the
answer cannot be given solely in terms of the minors of the matrix, though, as the
cokernel of the transpose matrix may have characteristics completely different
from the original.
So, to be uncompromisingly candid, our baby steps do not begin to address
some of the most elementary yet fundamental problems of commutative algebra.
While this may be taken by some to be disheartening, to say the least, we
can (and the authors do) take the view that there is a lot of growing yet to
do, and perhaps the recipe resides somewhere in the willingness to accept that
our accomplishments are as yet meager, but we should not lose sight of the
fundamental questions.
This page intentionally left blank
V
EXACTNESS CRITERIA AT WORK
Since the time that the structure theorems of the last chapter appeared, there
have been any number of applications and refinements of these results. What we
have tried to do is to select some area in which these theorems have played an
essential role, and in which there are still lurking some fascinating, only partially
explored areas of investigation.
We have chosen to look at resolutions of Gorenstein ideals of depth three
in a local ring, and their powers. (Recall [Definition II.4.10] that the depth
of an ideal, also called its grade, is the length of a maximal R-sequence in
I.) A Gorenstein ideal is a perfect ideal, I, of grade g, with the property
that TorR ∼
g (R/I, k) = R, where k is the residue field of the local ring, R. For
example, any complete intersection (i.e. an ideal generated by an R-sequence)
is a Gorenstein ideal, as the Koszul complex associated to any minimal set
of its generators is a minimal resolution and the last term of that complex is
precisely R.
We will look at Gorenstein ideals of depth three for a number of reasons. First,
in much the same spirit of the Hilbert–Burch Theorem (cf. Example IV.2.5), it is
possible to describe such ideals in terms of the minors of a certain kind of matrix,
namely an alternating one. We will show that all of these ideals are generated
by the so-called Pfaffians of order n − 1 of an alternating matrix of odd order,
n. Another reason is that for the first time in this book, the algebra structure
of such a resolution comes into play. It still is not completely clear just how
large a role the algebra structures of finite free resolutions must play in their
study, but these examples, as well as the vast amount of work on Poincaré series,
would indicate that such structure should not be ignored. The development in
the first section, where we deal exclusively with the Gorenstein ideal of grade
three follows to a large extent the treatment in Reference [25].
In Section V.2, we build on the previous one and construct resolutions for
all powers of the ideals in Section V.1. While this entails a bit of additional
technique, the results exhibit a phenomenon which seems to occur with some
frequency: a growth of homological dimension up to a critical point, and then
a constant value from a certain point on. This class of ideals also provides us
with the opportunity to illustrate the use of a counting method from represent-
ation theory to provide us with an idea of what the resolutions of such ideals
may look like. An elaboration of the material of Section V.2 can be found in
Reference [13].
128 Exactness criteria at work
0 → R → R → R/I → 0.
0 → R → R2 → R → R/I → 0.
The condition that the last term of the minimal resolution be free of rank one
(the condition that we placed on Tor) does not provide much wiggle room for the
resolution. But now our Hilbert–Burch theorem tells us that, since the depth of
I is two, I is generated by a regular sequence of length two, and our resolution
is essentially the Koszul complex associated to those generators.
Thus, we see that Gorenstein ideals of depth one or two must be generated by
regular sequences. This is why the real investigation of Gorenstein ideals has to
begin with depth three. Of course, one might be led to suspect that Gorenstein
ideals are all complete intersections; what follows should disabuse the reader of
that idea.
V.1.1 Pfaffians
Let R be a noetherian commutative ring, and F a finitely generated free
R-module. A morphism f : F ∗ → F is called alternating if, with respect
to some (and therefore every) basis and dual basis of F and F ∗ , the matrix
W = (wij ) of f satisfies wij + wji = 0 whenever i = j, and wii = 0 for every i.
Using the isomorphism
HomR (F ∗ , F ) ∼
= (F ∗ )∗ ⊗ F ∼
= F ⊗ F,
Proof Use induction on t and the formula in Lemma III.3.2 (cf. Reference [25],
pp. 458–460). 2
Corollary V.1.2 If rankF = 2k, then using the isomorphism
HomR (Λ2k F ∗ , Λ2k F ) ∼
= Λ2k F ⊗ Λ2k F,
the element Λ2k f corresponds to ϕ(k) ⊗ ϕ(k) .
Proof Set t = 2k in (V.1.1). Then the only non-zero summand is
(−1)k ϕ(k) (a) ϕ(k) ,
/ 0
since rankF = 2k. But ϕ(k) (a) = (−1)k ϕ(k) , a , where −, − stands for the
action of (ΛF ∗ )∗ on ΛF ∗ and we identify ϕ(k) ∈ Λ2k F with its image under the
canonical map
Λ2k F → Λ2k (F ∗ )∗ → (Λ2k F ∗ )∗ .
/ 0
It follows that Λ2k f (a) = ϕ(k) , a ϕ(k) for every generator a of Λ2k F ∗ . 2
Corollary V.1.2 has the following meaning. Choose an orientation of F , that
is, an isomorphism Λ2k F ∼ = R. Then Λ2k f corresponds to the determinant of the
appropriate matrix of f , and such a determinant is the square of the element
of R corresponding to ϕ(k) , called the pfaffian of f . We write Pf(f ), and also
Pf(W ), if W is a matrix associated with f .
If we take any alternating submatrix (of even order) W of W , we may consider
Pf(W ). It is the analog of the minor of a determinant. The following lemma
expresses a pfaffian of f of order 2m in terms of pfaffians of order 2m − 2 (rankF
is unspecified).
Lemma V.1.3 For every a ∈ Λ2m+1 F ∗ , we have:
a(ϕ(m+1) ) = ϕ(m) (a)(ϕ).
130 Exactness criteria at work
where
1 if i < i0
γii0 = 0 if i = i0
−1 if i > i0 ,
and Pf(W )h1 ,...,hr denotes the pfaffian of the alternating submatrix of W formed
by deleting rows and columns indexed by h1 , . . . , hr .
For every fixed i0 and j0 such that i0 = j0 ,
m
(−1)i+i0 +1 γii0 wij0 Pf(W )ii0 = 0.
i=1
and as b ranges over a basis of F , b(e) ranges over a basis of Λ2k F . Hence im(g)
is the ideal generated by the coordinates of ϕ(k) , that is, im(g) = P f2k (f ).
We first check that E is a complex. Since f ∗ = −f , it is enough to show
gf = 0. Now, for a ∈ F ∗ ,
gf (a) = ϕ(k) (e)(f (a)) = ϕ(k) (e)(a(ϕ)) = (ϕ(k) (e) ∧ a)(ϕ)
= −(a ∧ ϕ(k) (e))(ϕ) = −a(ϕ(k) (e)(ϕ)).
But by Lemma V.1.3,
ϕ(k) (e)(ϕ) = e(ϕ(k+1) ) = 0,
where
0 if k is odd
Tk = Λ2 Pk/2 if k is of the form 4n + 2
S2 Pk/2 if k is of the form 4n.
Since S2 (P) is isomorphic to P in degrees 0 and 1, the comparison theorem for
projective complexes says that there exists a map of complexes ϕ : S2 (P) → P
which extends the natural isomorphism R/a ⊗R R/a → R/a, and which is the
identity on the subcomplex R ⊗R P ⊆ S2 (P). If we set
f · g = ϕ(f ⊗ g),
↓ sg−1 ↓ s2 ↓ s1
d∗ d∗ d∗
F∗ : 0 → R →
1
F1∗ ∗
→ · · · → Fg−2 ∗
→ Fg−1 → Fg∗ .
g−1 g
and write (wij ) for the matrix of d2 relative to these bases. Thus, writing −, −
for the pairing between F2 and F1 ,
wij = εi , d2 (εj ) = εi · d2 (εj ) ∈ R = F3 ,
where · is the product for the algebra structure on F. We must show that wij =
−wji and wii = 0. Applying d3 to the above equation, and using the formula for
the differentiation of a product, we get
d3 (wij ) = d2 (εi ) · d2 (εj ) + εi · d1 (d2 (εj ))
= d2 (εi ) · d2 (εj ).
Since · satisfies the graded commutative law and d2 (εi ) and d2 (εj ) have degree
1, we have
d3 (wij ) = −d3 (wji ), d3 (wii ) = 0.
Since d3 is a monomorphism, this shows that d2 is alternating.
We next show that d2 satisfies the other conditions of Part 1 of Theorem V.1.8.
By Theorem IV.1.7, rank d2 = n − 1 and In−1 (d2 ) has grade ≥ 2. By statement
3 of Corollary V.1.6, n − 1 is even, so n is odd. Since F is a minimal resolution,
im(d2 ) ⊆ mF1 . To see that grade P fn−1 (d2 ) ≥ 3, we first note that by statement
1 of Corollary V.1.6 , rad P fn−1 (d2 ) = rad In−1 (d2 ), so it is enough to show that
grade In−1 (d2 ) ≥ 3. Since In−1 (d2 ) = I(d2 ), and I(d1 ) = I1 (d1 ) = a has grade 3
by assumption, we are done by Part (a) of Proposition IV.1.11.
Having proved that d2 satisfies all the conditions of the map f in Part 1 of
Theorem V.1.8, proceeding as in Subsection V.2.2, we can construct an exact
sequence
g∗ g
E : 0 → R → F1∗ →2 F1 → R.
d
But under our identifications F3 = R and F2 = F1∗ , we have d3 = d∗1 . Since both
d∗1 and g ∗ are kernels of d2 , there is a unit u ∈ R such that the following diagram
is commutative:
g∗
R −→ F1∗
↓u .
∗
R −→
∗
F 1
d1
Hence ug = d1 and
P fn−1 (d2 ) = im(g) = im(ug) = imd1 = a,
as required.
This completes the proof of Theorem V.1.8.
One should remark that, given a Gorenstein ideal, a, of grade 3 as above,
minimally generated by n = 2k + 1 elements, we can rewrite the resolution
g∗ g
E : 0 → R → F1∗ →2 F1 → R
d
136 Exactness criteria at work
as follows:
g f g
E : 0 → Λn F1 → Λn−1 F1 → F1 → R,
by means of the isomorphisms Λk F → Λn−k F ∗ such that a → a(e) [e = chosen
orientation of Λn F ∗ ]. Then the pairing on E , which gives the duality expressed
by Theorem V.1.10, is precisely the multiplication
F1 ⊗R Λn−1 F1 → Λn F1 , R ⊗R Λn F1 → Λn F1
in the exterior algebra.
Finally, we observe that there is a close analogy between E and the Koszul
complex (which E must be, if n = 3). Namely, we note that the following
diagram is commutative, where the first row is a truncation of the Koszul complex
associated with g : F1 → R, where γ : ΛF1 → ΛF1 is the map a → a ∧ ϕ(k−1) ,
and ϕ is the element of Λ2 F1 corresponding to f .
g
Λ3 F1 → Λ2 F1 → F1 → R
↓γ ↓γ
g f g
Λn F1 → Λn−1 F1 → F1 → R.
for it always has grade 3 while its homological dimension depends on m (in a
very nice way).
where m denotes multiplication in the appropriate algebras and CF0 stands for
the element of
F0 ⊗R F ∗ ∼
= HomR (F0 , F0 )
0
F0 −→ F0 ⊗R F0 ⊗R F0∗ −→ Λ2 F0 ⊗R F0∗ ,
1⊗CF0 m⊗1
sends each ei to j (ei ∧ ej ) ⊗ εj . Hence (ei ∧ ej ) is the matrix associated to
f with respect to the bases of F and F ∗ induced by those of F0 and F0∗ (we
denote the induced bases with the same symbols used for the original ones). By
means of the identification ei ∧ ej → Xij , the matrix (ei ∧ ej ) coincides with the
generic alternating matrix X introduced at the beginning of this section, and
A = S(Λ2 F0 ) is isomorphic to S = R[X].
From now on, we use S to indicate both R[X] and S(Λ2 F0 ), dropping the
symbol A.
The morphism, f , is called the generic alternating map. Asseen in Section
1, f corresponds to an element α ∈ Λ2 F ∗ , namely, we have α = i<j Xij εi ∧ εj .
Moreover, the grade 3 ideal P f2k (X), which we denote by a, for short, has the
by now familiar resolution (a resolution of S/a, i.e.):
g∗ f g
E : 0 → S → F → F ∗ → S.
The morphism f is as above. As for g, choose an orientation e ∈ Λn F , that is,
an identification between Λ2k F and F ∗ . Then g can be identified with
ω2k : Λ2k F → S, a → β(α(k) ⊗ a),
where β stands for the natural pairing Λ2k F ∗ ⊗S Λ2k F → S given by the ΛF ∗ -
structure of ΛF .
Concretely, g can be viewed as sending each εi to Ti (X) = (−1)i+1 Pf(X)i .
Finally, g ∗ (1) = α(k) .
138 Exactness criteria at work
(λ1 , 1, . . . , 1) = (λ1 , 1t ),
% &# $
t
i1 ··· iλ1
j1
.. .
.
jt
• A basis of L(λ1 ,1t ) F is given by all tableaux such that the indices in the first
row are strictly increasing and those in the first column are weakly increasing
(“standard tableaux”).
• A tableau which is not standard is equal to a Z-linear combination of
standard tableaux (“straightening law”).
Powers of pfaffian ideals 139
where (µ1 , 1u ) ranges over all the hooks which are nested in (λ1 , 1t ). But as
GL(F1 ) × GL(F2 )-modules, such an isomorphism usually holds only up to a
filtration.
where (ω2k )0 is the R-map Λ2k F0 → Sk (Λ2 F0 ) inducing ω2k over S. Since
S2 (Λ2k F0 ) ∼
= S2 F0∗ = L(2k+1,1,1) F0∗ ∼
= L(2k,2k) F0 , ϑ1 is in fact a morphism
where l αl1 ⊗ αl1 is the image of α given by the component Λ2 F ∗ → F ∗ ⊗S F ∗
of the diagonal map, and where p(2k+1,1) is the projection Λ2k+1 F ∗ ⊗S S1 F ∗ →
L(2k+1,1) F ∗ , one conjectures that ϑ2 and ϑ3 are induced by
Λ2k F ∗ ⊗S S1 F ∗ → Λ2k+1 F ∗ ⊗S S2 F ∗ , u ⊗ v →
αl1 ∧ u ⊗ vαl1
l
and by
Λ2k−1 F ∗ → Λ2k F ∗ ⊗S S1 F ∗ , u →
αl1 ∧ u ⊗ αl1 ,
l
respectively. One checks that ϑ2 and ϑ3 are actually well defined, by invoking
the following more general result, which holds for every ring R.
Lemma
V.2.1 Let ϕ be the map Λa F ∗ ⊗S Sb F ∗ → Λa+1 F ∗ ⊗S Sb+1 F ∗ , u⊗v →
α
l l1 ∧ u ⊗ vαl1 . Then ϕ induces a morphism
ϕ : L(a,1b ) F ∗ → L(a+1,1b+1 ) F ∗ .
Proof Recall the expression of a hook Schur module
as a cokernel, and perform
calculations applying the diagonal map to α = i<j Xij εi ∧εj (cf. Reference [13],
p. 472). 2
It is not hard to see that our candidate is indeed a complex. That ϑ2 ◦ ϑ3 = 0
can be seen as follows. Let ϕ 2 and ϕ 3 be the maps inducing ϑ2 and ϑ3 ,
respectively. Then
2 (ϕ
ϕ 2
3 (u)) = ϕ Xij (εi ∧ u ⊗ εj − εj ∧ u ⊗ εi )
i<j
= Xij Xpq (εp ∧ εi ∧ u ⊗ εq εj − εp ∧ εj ∧ u ⊗ εq εi
i<j p<q
− εq ∧ εi ∧ u ⊗ εp εj + εq ∧ εj ∧ u ⊗ εp εj )
and the double sum equals zero (think of what happens if one interchanges
j ↔ q and i ↔ p). We also have that ϑ1 ◦ ϑ2 = 0 because if εi stands for
ε1 ∧ · · · ∧ εi−1 ∧ εi+1 ∧ · · · ∧ εn , then
ϑ1 (ϑ2 (p(2k,1) (εi ⊗ εj ))) = Xpq ϑ1 (p(2k+1,1,1) (εp ∧ εi ⊗ εj εq ))
p=q
= Xiq (−1)i−1 Tj (X)Tq (X)
q
1 2
i−1
= (−1) Tj (X) Xiq Tq (X)
q
and the sum in brackets is zero by the last formula of Remark V.1.4.
This concludes our heuristic considerations.
Powers of pfaffian ideals 143
(n = 2k − 1, as above).
146 Exactness criteria at work
for some t ≤ t and some non-negative integer j. To Xp we assign S and all the
summands of L(a,1b ) (H ∗ ⊕ G∗ ) having µ1 = n − j and u = 2t − j for some t ≤ p
and some non-negative integer j.
It is easy to see that Xp = M1 (cf. Reference [13], proposition at the end of
p. 481). Recalling the straightening law at the end of Subsection V.2.2, it is not
hard to check that each Xt , 0 ≤ t ≤ p, is indeed a complex (cf. Reference [13],
proposition on p. 482).
We now describe the factors Xt /Xt−1 , 0 ≤ t ≤ p − 1 (X−1 is set equal to 0),
and prove their exactness.
The modules occurring in Xt /Xt−1 are given by µ1 = n − j and u = 2t − j,
with j ranging between 0 and q, where
2t if 2t < n
q=
n − 1 if 2t > n .
is defined by
p(µ1 ,1u ) (x ⊗ y) ⊗ z1 ⊗ z2
→ p(µ1 +1,1u+1 ) ((αH ∗ )l1 ∧ x ⊗ y(αH ∗ )l1 ) ⊗ z1 ⊗ z2 .
l
is defined by
x ⊗ y ⊗ z → (−1)µ1 x ⊗ y(αG∗ )l1 ⊗ (αG∗ )l1 ∧ z.
l
ψ ⊗S S , namely (recall Λn H ∗ ∼
= S ), the morphism defined by
(ε )b3 · · · (εn )bn → T3 (X )b3 · · · Tn (X )bn .
% 3 &# $
b3 +···+bn =m
But Cm
is a complex of the same type of Cm , relative to the generic alternating
matrix X of order 2(k − 1) + 1. Therefore it is exact by induction hypothesis.
This concludes our discussion of the exactness of the complexes Cm .
VI
WEYL AND SCHUR MODULES
.
Shape matrices and tableaux 151
Note that the diagram ignores the fact that the left-most column of the shape
matrix consists completely of zeroes. Also notice that the empty diagram
corresponds to the zero matrix.
The shapes illustrated above are those that we will have most to do with, but
clearly we can associate to any shape matrix, A = (aij ), a Ferrers diagram, or
simply diagram: we just set up a grid equal to the effective size of the matrix
(say, s × t), and throw away the boxes whose entries are equal to zero. That is,
the (i, j)th box lies in the diagram if and only if aij = 1. If A is the shape matrix,
we will sometimes denote by (A) its corresponding diagram.
The partition shape has a uniquely associated partition, namely the sequence
λ = (λ1 , . . . , λs , . . .) where λi is the non-negative integer equal to the number
of ones in row i. Clearly, the sequence is decreasing: λ1 ≥ λ2 ≥ · · · ≥ λs ≥ · · · .
We say the length of λ is s if s is the smallest non-negative integer such that
λs+t = 0 for every positive integer t.
The skew-shape has two partitions uniquely associated to it, namely, λ =
(λ1 , . . . , λs , . . .) and µ = (µ1 , . . . , µt , . . .), with µi ≤ λi for all i, and such that
the length of µ is strictly less than that of λ. One then thinks of the shape as
the result of removing from the shape of λ the subshape corresponding to µ. In
fact, the notation most often used for a skew-shape is λ/µ. We say the length
of λ/µ is the length of λ. If one removes the condition that the length of µ be
152 Weyl and Schur modules
strictly less than that of λ, then we have other pairs of partitions (λ , µ ) that
will yield the same diagram. In that case, we still use the same notation, λ /µ
(but the length of λ /µ stays equal to the length of the previous λ).
Finally, we see that the almost skew-shape would be a skew-shape but for its
last row, which, rather than projecting beyond (or flush with) the penultimate
row, does not make it out that far to the left. In short, it would be a skew-
partition but for that inadequacy in the last row.
In our examples above, the partition λ associated to (P ) is (4, 3, 3, 2); the
pair of partitions associated to (S) are λ = (5, 4, 3, 2) and µ = (2, 2, 1). (For
convenience we have eliminated the zeroes to the right in our notation.) For (AS),
we might take the skew-partition to be (7, 7, 6, 5)/(3, 2, 1) with a bar having the
entries (1, 1) in the fourth row, or we might take (7, 7, 6, 5)/(3, 2, 1, 1), with a bar
having entries (0, 1) in the fourth row.
Another way to denote an almost skew-shape, which closely parallels the nota-
tion for a skew-shape is to first define an almost partition to be a sequence
(µ1 , . . . , µn ) such that µ1 ≥ · · · ≥ µn−1 and 0 ≤ µn ≤ µ1 . We then can denote
(not necessarily uniquely) an almost skew-shape by λ/µ, where λ is a partition
of length n, and µ is an almost partition having exactly n terms and satisfying
µi ≤ λi for all i. We can go one step further, and say that an almost partition,
µ, is of type n − (i + 1) if i is the largest integer less than n such that µn ≤ µi ,
and we say that the type of the almost skew-shape λ/µ is equal to the type of
µ. Clearly, the type is independent of the choice of λ and µ used to describe the
almost skew-shape. The choice of the pair (λ, µ) can be made canonical if, in the
case of type zero, we choose µn = 0, while for type greater than zero, we choose
µn−1 = 0. We define the length of the almost skew-shape to be the length of
the canonical partition, λ.
With this terminology, we see that an almost skew-shape of type 0 is a
skew-shape, and that for almost skew-shapes of length n, we can have types
0, 1, . . . , n− 2. In particular, almost skew-shapes of length 2 are necessarily skew-
shapes; for length 3, there are only skew-shapes and almost skew-shapes of type
1, and so on.
We spoke of shape matrices as infinite matrices in order not to have to specify
last row or column when we talked about subshapes. However, we see that a
shape matrix whose last row is row s and whose last column is column t, can
be thought of as an s × t-matrix; when we draw diagrams or shapes, we will
generally avoid the dots that we were forced to put into the illustrations above.
If we have two shape matrices A and B, with B ⊆ A, we will assume that they
are both s × t-matrices, simply by augmenting where necessary by zeroes.
is also
1. If A is a partition (or skew-partition) matrix, then its transpose, A,
a partition (or skew-partition) matrix.
Shape matrices and tableaux 153
The notation and terminology used above differ slightly from those used else-
where. For instance, in I. Macdonald’s book [66], the transpose is called the
“conjugate,” and is denoted by A ; in Reference [3], it is called the “dual,” but
denoted in the way it is above.
VI.1.2 Tableaux
Definition VI.1.4 Let A be a shape matrix and S a set. A tableau, T , of
shape A with values in S is a filling-in of the diagram (A) by elements of S.
We denote the tableau by the ordered pair T = ((A); τ ), where τ is the filling-in
of (A) by S.
We could have said that τ is a function from (A) to S, but for the fact that
we have not given a formal enough definition of “diagram” to do this. But if
one regards the diagram as a collection of cells, then τ would be a function with
domain (A). In most cases, we will simply refer to the tableau as T , with the set
S an understood ordered basis of a finitely generated free module. This is how
we saw tableaux used in earlier chapters.
Occasionally we will use the term row tableau; this is simply a tableau the
diagram of whose shape consists of one row.
For further work with tableaux, it will be convenient to have a quasi order on
tableaux with values in a totally ordered set. Suppose, then, that S is a totally
ordered set, say S = {s1 , . . . , sn } with s1 < · · · < sn , and suppose T is a tableau
with values in S. Define Tij to be the number of elements in {s1 . . . , si } that
appear in at least one of the first j rows of the diagram of T . Now suppose that
T is another tableau.
Definition VI.1.5 We say that T ≤ T if Tij ≥ Tij for every i and j. We say
that T < T if T ≤ T and for some i, j we have Tij > Tij .
To see that this is a quasi order and not an order, consider our set S with
three elements: S = {s1 , s2 , s3 }, with s1 < s2 < s3 , and consider the diagram
154 Weyl and Schur modules
s1 s2 s2 s3
T = s2 s3
s3
and
s2 s3 s2 s1
T = s3 s2
s3
are such that T ≤ T and T ≤ T , but T and T are clearly not equal. However,
the “≤” relation is both reflexive and transitive, as can easily be checked.
powers of F into the indicated tensor products. We define the maps µa : Λa1 F ⊗
· · · ⊗ Λal F → Λα F and µa : Sa1 F ⊗ · · · ⊗ Sal F → Sα F to be the multiplication
maps from the indicated tensor products of exterior and symmetric powers to
the indicated exterior and symmetric powers.
Definition VI.2.1 (Weyl and Schur maps) Let F be a free module over the
commutative ring, R. For the s×t shape matrix A = (aij ), set ri = (ai1 , . . . , ait ),
t s
cj = (a1j , . . . , asj ), ρi = j=1 aij , γj = i=1 aij . The Weyl map associated
to A, ωA , is the map
where, since all the aij are equal to zero or one, we have identified Daij F with
Λaij F for all i, j; the map θW is the isomorphism comprising all of these iden-
tifications together with rearrangement of the factors. Pictorially what we have
Modules associated to shape matrices 155
is the following:
Da11 F ⊗ · · · ⊗ Da1t F
⊗
Dρ1 F ⊗ · · · ⊗ Dρs F → .
.
θW
−→
.
⊗
Das1 F ⊗ · · · ⊗ Dast F
a
Λ 11 F ⊗ · · · ⊗ Λa1t F
⊗
.. → Λγ1 F ⊗ · · · ⊗ Λγt F.
.
⊗
as1
Λ F ⊗ · · · ⊗ Λast F
The Schur map associated to A, σA , is the map
σA : Λρ1 F ⊗ · · · ⊗ Λρs F → Sγ1 F ⊗ · · · ⊗ Sγt F
defined as the composition
σA = µc1 ⊗ · · · ⊗ µct θS δr1 ⊗ · · · ⊗ δrs
where, since all the aij are zero or one, we have identified Λaij F with Saij F
for all i, j; the map θS is the isomorphism comprising all of these identifications
together with rearrangement of the factors. We can view the definition of the
Schur map “pictorially” in the same way we did the Weyl map.
Definition VI.2.2 (Weyl and Schur modules) Let F be a free R-
module, and A a shape matrix. We define the Weyl module of F associated
to A, denoted KA F , to be the image of ωA . We define the Schur module of
F associated to A, denoted LA F, to be the image of σA .
Remark VI.2.3 The following observations are easy to check and very
useful.
1. If A is a non-zero shape matrix with its initial column consisting only of
zeros, and B is the shape matrix with that initial column removed, it is clear
that the Weyl and Schur maps associated to A and B are the same. Hence,
we will generally assume that our shape matrices have at least one entry in
the first column equal to one.
2. If A is a shape matrix, and B is the shape matrix obtained from A by
a permutation of its rows (columns), then the associated Weyl and Schur
modules of these matrices are isomorphic.
With the definition of Weyl and Schur modules to hand, a natural question to
consider is whether these modules are free over the ground ring and, if so, how
can we describe a basis. For example, if we take a one-rowed partition λ, what
156 Weyl and Schur modules
are the Weyl and Schur modules associated to it? The Weyl module is the image
of the map Dλ F →ωλ Λ1 F ⊗ · · · ⊗ Λ1 F , where ωλ is the diagonalization map.
% &# $
λ
Clearly, then, the image is isomorphic to Dλ F itself, that is, Kλ F = Dλ F . In a
similar way we can show that Lλ F = Λλ F . In both of these cases, the modules
are clearly free R-modules, and we have a very concrete description of bases for
them. Not only do we have explicit descriptions of their bases, we even have a
description in terms of tableaux. In the case of Dλ F , a basis is parametrizable by
the set of all one-rowed tableaux: xi1 xi2 · · · xiλ where {x1 , . . . , xm } is an
ordered basis of the free module, F , and i1 ≤ · · · ≤ iλ . In the case of Λλ F , we have
that a basis is parametrizable by all one-rowed tableaux: xi1 xi2 · · · xiλ
where i1 < · · · < iλ .
It will be proven later that if λ/µ is an almost skew-partition, then Kλ/µ and
Lλ/µ are free, and bases for them can be parametrized by certain sets of tableaux
of shape λ/µ satisfying combinatorial conditions.
if we take D(F ⊗ Rn ) ∼
= D(F ⊕ · · · ⊕ F ), we see that D(F ⊗ Rn ) is equal to
% &# $
n
D(F ) ⊗ · · · ⊗ D(F ). This is natural with respect to the action of GL(F ), but
% &# $
n times
clearly not with respect to the action of GL(Rn ). In fact, we moved to the
notation Rn rather than G to indicate that we have made a choice of basis in
our free module, G. We can, though, use G in our preliminary discussion and,
assuming that the rank of this free module is n, still see that “in some way”,
D(F ⊗ G) ∼= D(F ) ⊗ · · · ⊗ D(F ). Now we want to introduce convenient notation
% &# $
n
to exhibit this isomorphism, as well as to get to the letter-place conventions.
To this end, let us suppose that G has the (ordered) basis, {y1 , . . . , yn } with
y1 < · · · < yn , and for any x ∈ D1 (F ), let us denote by (x|yi ) the element x ⊗ yi ,
(k)
and by (x(k) |yi ) the element corresponding in D(F ) ⊗ · · · ⊗ D(F ) to (x|yi )(k) ,
% &# $
n
that is, to the element in that n-fold tensor product of D(F ) having x(k) in the
ith factor.
• The picture to keep in mind is: (x|yi ) is the element 1 ⊗ · · · ⊗ x ⊗ · · · ⊗ 1.
% &# $
i
Now
k!(x ⊗ yi )(k) = (x ⊗ yi )k = (1 ⊗ · · · ⊗ x ⊗ · · · ⊗ 1)k
% &# $
i
= 1 ⊗ · · · ⊗ x ⊗ · · · ⊗ 1 = k! (1 ⊗ · · · ⊗ x(k) ⊗ · · · ⊗ 1),
k
% &# $ % &# $
i i
(k)
so that the above definition of (x(k) |yi ) makes sense.
Finally, if l = l1 + · · · + ln , and x ∈ Dl (F ), we set
(l )
(l )
(x|y1 1 · · · yn(ln ) ) = (x(l1 )|y1 1 ) · · · (x(ln )|yn(ln ) ),
where x(l1 )⊗· · ·⊗x(ln ) indicates the image of the diagonal map into Dl1 (F )⊗
· · · ⊗ Dln (F ) applied to our element x.
Remark VI.3.1 The identities and conventions that we adopt for our discus-
sion are those that are clearly valid if one works over the ring of integers (as is
the case illustrated above, where we have cancelled k! because there is no torsion
over the integers). We will continue to do this in our treatment of the letter-
place algebra and all other structures that are transportable from Z to arbitrary
commutative base rings.
A simple illustration, just to fix our ideas, is this: Suppose x1 , x2 and x3 are
in D1 (F ), and consider the element
(2) (2)
(x1 x2 x3 |y1 y2 y3 ) ∈ D2 (F ) ⊗ D1 (F ) ⊗ D1 (F ).
158 Weyl and Schur modules
(2) (2)
(x1 x2 |y1 )(x2 |y2 )(x3 |y3 ) + (x1 x2 |y1 )(x3 |y2 )(x2 |y3 )+
(2) (2) (2)
(x1 x3 |y1 )(x2 |y2 )(x2 |y3 ) + (x2 |y1 )(x1 |y2 )(x3 |y3 )+
(2) (2) (2)
(x2 |y1 )(x3 |y2 )(x1 |y3 ) + (x2 x3 |y1 )(x1 |y2 )(x2 |y3 )+
(2)
(x2 x3 |y1 )(x2 |y2 )(x1 |y3 ).
(a ) (a )
nthe symbol (w|y1 · · · yn ) equal to zero if the degree
1 n
(♠) We agree to set
of w is not equal to i=1 ai . The element w is supposed to be a homogeneous
element of D(F ).
The letter-place notation we have been using above lends itself very naturally
to writing tableaux. That is, suppose we wanted to write the product of the
(2) (2) (2)
above element, (x1 x2 x3 |y1 y2 y3 ) with, say, (x3 x1 |y1 y2 y3 ). As we saw above,
each of these terms is a sum of a number of addends, so that the notation we
have for each of these terms is already of considerable convenience. But now,
instead of using juxtaposition to denote the product of these two terms, let us
use “double tableau” notation, that is, let us write
(2) (2)
x1 x2 x3 y1 y2 y3
(2)
x3 x1 y1 y2 y3
We point this out to indicate that we may always assume that our tableaux
are given in such a way that in each row, the elements are increasing. The
terminology for this is that the tableaux are row-standard (a notion that we
have already encountered in Chapter III). Recall that in Chapter III, we wrote
out the rows of the tableau repeating letters instead of using divided powers. This
helps to talk about the columns of a tableau; for instance, the tableau above has
two rows and four columns (the number refers to the arrays in the letters as well
as the places).
Usually, we call the basis of F the letters, while the basis of G is called the
places. A basic word of degree k is simply a basis element of Dk (F ), while
a word of degree k is a linear combination of basic words of degree k. Usually
Letter-place algebra 159
The proof breaks up into two parts: the double tableaux of type (G), with
l ajl = kj , generate, and the number of such tableaux is equal to the number
of tableaux of type (W ) above, for fixed k1 , . . . , kn . The first part is given in
Appendix A, Section 1, and the second part in Appendix A, Section 2.
(a1 a2 |p1 ∧ p2 ) = (a1 |p1 )(a2 |p2 ) + (a2 |p1 )(a1 |p2 ).
the basis of F , and the left side consists of words in the negative alphabet of
places, the basis {1, . . . , n} of Rn .
To see that our usual basis elements of Λk1 F ⊗ · · · ⊗ Λkn F can be expressed
as double tableaux, consider the following example.
• The element x2 ∧ x3 ∧ x5 ⊗ x1 ∧ x3 ⊗ x2 ∧ x4 ⊗ x3 ∧ x5 ∧ x6 ∈ Λ3 F ⊗ Λ2 F ⊗
Λ2 F ⊗ Λ3 F can be expressed as the double tableau:
(2)
x2 1 3
(3)
x3 1 2 4
(2)
x 1 4
5
x1 2
x4 3
x6 4
On the right hand side of the tableau we have omitted the wedge, and simply
spread the basis elements out along the row. We used the divided power
notation on the left hand of the column to simplify writing. Really, the top
row of the tableau above should look like:
(x2 x2 |1 3).
where we are again cheating a bit in that we are writing the left hand side in
terms of divided powers rather than as spread-out sequences. Just to make sure
that we understand what all the notation means, we remark that the tableau
above is equal to:
E = x3 ∧ x5 ⊗ x3 ⊗ x4 ∧ x5 ∧ x6
+ x3 ∧ x5 ⊗ x4 ⊗ x3 ∧ x5 ∧ x6
+ x4 ∧ x5 ⊗ x3 ⊗ x3 ∧ x5 ∧ x6 .
In the same way the double standard tableaux generate the tensor product of
divided powers, these double standard tableaux generate the tensor product of
exterior powers. We have the following theorem.
Theorem VI.3.3 The set of standard double tableaux having the ith place
counted ki times is a basis for Λk1 F ⊗ · · · ⊗ Λkn F .
(F I) (w|p) = (x1 |qσ(1) ) · · · (xk |qσ(k) ),
σ
The individual factors of the tensor product above have been described in
detail; the product of two terms from different components of the product
is simply the tensor product of these terms, while the product (l1 |p)(l2 |p) =
(−1)|(l1 |p)||(l2 |p)| (l2 |p)(l1 |p).
2{(l1 |s)(l2 |s)}, so that for q and s positive places, it makes sense to talk about
(k)
the higher divided powers of the place polarizations, ∂s,q , namely ∂s,q . In the
case of the divided square just discussed, for instance, we see that the equation
(2)
may be interpreted as ∂s,q (l1 l2 |q (2) ) = (l1 l2 |s(2) ). In general, then, we have
(k)
∂s,q (w|q (m) ) = (w|q (m−k) s(k) ),
where q and s are positive places.
One fundamental identity, which is easy to prove, is the following.
Fact 1 Let p, q, r be places with q and p positive, and consider the place
polarizations ∂r,q , ∂q,p and ∂r,p . Then
∂r,p = ∂r,q ∂q,p − ∂q,p ∂r,q .
166 Weyl and Schur modules
∂r,p (w|p(a) q (b) r(c) . . .) = (w|p(a−1) q (b) rr(c) . . .) = (c + 1)(w|p(a−1) q (b) r(c+1) . . .).
But we have
and
(Cap ) (b) (a)
∂q,p ∂r,q = (a−k) (b−k) (k)
(−1)k ∂r,q ∂q,p ∂r,p .
k≥0
Proof For a = b = 1, this is just the commutation rule (for both Cap and
Cap ). One can now proceed, by direct calculation, or we can employ a standard
trick and prove this by induction, first on a for b = 1, and then for all a and
b. The standard trick is to behave as though we were working over the field of
rational numbers (which allows us to divide by integers), and then assert that
since our proposition works there, it works in general (see Remark VI.3.1).
We will therefore proceed to prove our proposition by induction. As we pointed
out, our proposition is true for a = b = 1, so we may let b = 1 and assume the
Weyl and Schur maps revisited 167
(Cap )− (b)
∂q,p (b)
∂r,q = ∂r,q ∂q,p − ∂q,p
(b−1)
∂r,p .
For the Weyl map, we are going to consider the basis, L+ , of F as a positive
letter alphabet (in the letter-place language), and our place alphabet P = P + "
P − , where P + = {1, . . . , n} and P − = {1, . . . , m}. For the Schur map, we are
going to regard the basis of F as a negatively signed letter alphabet, L− , and
our place alphabet the same as the above.
We next observe that S(L+ |P) = D(F ⊗ Rn ) ⊗ Λ(F ⊗ Rm ), which contains
the subalgebras S(L+ |P + ) = D(F ⊗ Rn ) and S(L+ |P − ) = Λ(F ⊗ Rm ). Our
discussion of the letter-place algebra tells us that D(F ⊗ Rn ) = DF ) ⊗ · · · ⊗ DF
% &# $
n
while Λ(F ⊗ Rm ) = ΛF ⊗ · · · ⊗ ΛF . A similar discussion applies to the algebra
% &# $
m
S(L− |P) = Λ(F ⊗ Rn ) ⊗ S(F ⊗ Rm ).
What we will show is that our Weyl (or Schur) maps are compositions of place
polarizations that take us from our desired domain to our desired target through
S(L+ |P) (or S(L− |P)).
Although we can carry out this project for arbitrary shapes, we will restrict
ourselves to the class of shapes we have already discussed in Section VI.1, namely
almost skew-shapes. Recall that an almost skew-shape can be represented as
λ/µ where λ is a partition and µ is an almost partition. In order to conform
to the notation used to describe the shape matrix, A, above, we will assume
that our partition λ has length n, and that λ1 − µn = m if µ is a partition,
and that λ1 − µn−1 = m if µ is not a partition. A quicker way to say this is
that λ1 − min(µn , µn−1 ) = m. As we have noted before, we may as well set
min(µn , µn−1 ) = 0.
Using this notation for our shapes, we see that the numbers pi and γj above
become:
pi = λi − µi ; γj = λ̃j − µ̃j ,
∆λ/µ = ∆n · · · ∆1 .
is a map
∆λ/µ : DF ⊗ · · · ⊗ DF → ΛF ⊗ · · · ⊗ ΛF .
% &# $ % &# $
n m
i i (t +k )
which, on each summand, is equal to ∂i+1,i . Now define
Rel(λ/µ) = Dp1 ⊗ · · · ⊗ Dpi +ti +ki ⊗ Dpi+1 −ti −ki ⊗ · · · ⊗ Dpn ,
i ki
where the sum is taken over i = 1, . . . , n − 1, and all positive ki . And now define
λ/µ : Rel(λ/µ) → Dp1 ⊗ · · · ⊗ Dpn
to be the map which, for each i, is the map λ/µ,i .
Some kernel elements of Weyl and Schur maps 171
→ Dp1 ⊗ · · · ⊗ Dpn
(k )
to be the map which on each component is the place polarization ∂n,jj , and for
i = n − (τ + 1), define
s
#λ/µ,i : Dp1 ⊗ · · · ⊗ Dpi+ti−s+k ⊗ Dpi+1 ⊗ · · · ⊗ Dpn−1 ⊗ Dpn−(ti−s)−k
k=1
→ Dp1 ⊗ · · · ⊗ Dpn
to be, again, the map which on each component is the place polarization
(t −s+k)
∂n,ii .
We next define, for an almost skew-shape, λ/µ of type τ > 0, the overall
relations Rel(λ/µ) = Rel(p1 , . . . , pn ; t1 , . . . , tn−1 ) by:
Rel(p1 , . . . , pn ; t1 , . . . , tn−2 , 0)
n−2
tj
Dp1 ⊗ · · · ⊗ Dpj +kj ⊗ Dpj+1 ⊗ · · · ⊗ Dpn−1 ⊗ Dpn −kj
j=i+1 kj =1
s
Dp1 ⊗ · · · ⊗ Dpi +ti −s+k ⊗ Dpi+1 ⊗ · · · ⊗ Dpn−1 ⊗ Dpn −(ti −s)−k ,
k=1
and the map
λ/µ : Rel(p1 , . . . , pn ; t1 , . . . , tn−1 ) → Dp1 ⊗ · · · ⊗ Dpn
172 Weyl and Schur modules
Thus, the image of a polarization of high order from one place to another far
down the line is already in the image of one of the same order from, say, adjacent
places, plus the images, by lower order polarizations, between the original places.
This is what permits us to limit the index of summation.
The main focus of this section is the proof of the following essential result.
Theorem VI.6.1 Let λ/µ be any almost skew-shape. Then the composition
λ/µ ∆λ/µ
Rel(λ/µ) −→ Dp1 ⊗ · · · ⊗ Dpn −→ Λγ1 ⊗ · · · ⊗ Λγm
is zero. That is, the image of λ/µ is contained in the kernel of the Weyl map.
Proof The proof depends heavily on the Capelli identity (Cap) involving
positive-to-negative polarizations.
First assume that our shape is a skew-shape, that is, that τ = 0. We want to
show that on each summand
Dp1 ⊗ · · · ⊗ Dpi +ti +ki ⊗ Dpi+1 −ti −ki ⊗ · · · ⊗ Dn ,
(t +k )
i i
the map ∆λ/µ ∂i+1,i is zero. Since ∆λ/µ is a composition of compositions of
positive-to-negative place polarizations (recall that ∆λ/µ = ∆n · · · ∆1 ), we can
(ti +ki )
commute ∂i+1,i past all the maps ∆1 , . . . , ∆i , and thus we have to study what
Some kernel elements of Weyl and Schur maps 173
(t +k )
the composition ∆n · · · ∆i+1 ∂i+1,i
i i
∆i · · · ∆1 looks like. We further dissect this
i i (t +k )
composition, and focus on the parts surrounding ∂i+1,i , namely
(t +k )
∂λi+1 ,i+1 · · · ∂µi+1 +1,i+1 ∂i+1,i
i i
∂λi ,i · · · ∂µi +1,i .
Using recursion and the identity (Cap), one sees that
(t +k )
∂λi+1 ,i+1 · · · ∂µi+1 +1,i+1 ∂i+1,i
i i
pi+1
(t +k −α)
= ∂i+1,i
i i
∂λi+1 ,i+1 · · · ∂ˆβ1 ,i+1 · · · ∂ˆβα ,i+1 · · · ∂µi+1 +1,i+1 ∂β1 ,i · · · ∂βα ,i
α=0 βα <···<β1
where µi+1 < βα < · · · < β1 ≤ λi+1 , and the hat over the ∂ means that term is
to be omitted.
Next we observe that if α < ti , then each term
∂λi+1 ,i+1 · · · ∂ˆβ1 ,i+1 · · · ∂ˆβα ,i+1 · · · ∂µi+1 +1,i+1 ∂β1 ,i · · · ∂βα ,i
has at least pi+1 − ti polarizations from i + 1 to some negative place. But since
our original element has the place i + 1 only pi+1 − ti − ki times, our element is
carried to zero when α < ti . But if α ≥ ti , then the terms ∂β1 ,i · · · ∂βα ,i composed
with ∆i polarize the positive place i to α + pi ≥ ti + pi negative places between
µi+1 + 1 and λi . Since λi − (µi+1 + 1) = λi − µi + µi − µi+1 − 1 = pi + ti − 1, the
composition of these polarizations must take i to the same negative place twice,
which results in the zero map.
Now we assume that τ > 0, so that we must consider the extra summands
that comprise Rel(λ/µ). These are of two types:
D1 ⊗ · · · ⊗ Dpj +kj ⊗ Dpj+1 ⊗ · · · ⊗ Dpn−1 ⊗ Dpn −kj
and
D1 ⊗ · · · ⊗ Dpi +ti −s+k ⊗ Dpi+1 ⊗ · · · ⊗ Dpn−1 ⊗ Dpn −(ti −s)−k ,
To see what happens to terms of the first type under the composition, we again
(k )
see that the map λ/µ applied to them is simply ∂n,jj . We therefore just have
(k )
to look at ∆n ∂n,jj ∆n−1 . If we apply (Cap) as we did above, it is even easier to
see that the composition is zero, since the place j is sent to the negative places
µj + 1, . . . , λj , and µn > µj . For the term D1 ⊗ · · · ⊗ Dpi +ti −s+k ⊗ Dpi+1 ⊗
· · · ⊗ Dpn−1 ⊗ Dpn −(ti −s)−k , the same argument that was used above yields the
result. 2
Corollary VI.6.2 Let us define K̄λ/µ to be the cokernel of λ/µ . Then the
identity map on Dp1 ⊗ · · · ⊗ Dpn induces a map θλ/µ : K̄λ/µ → Kλ/µ .
Proof This follows immediately from the result above. 2
All of the above discussion carries over to the Schur map and Schur modules,
simply by replacing divided powers by exterior powers and exterior powers by
174 Weyl and Schur modules
symmetric powers. Or, if one wishes, one can simply replace the positive letter
alphabet by its negative counterpart. All the maps that we define are in terms
of the place alphabets, and these have not changed.
Proof It suffices to show that if the tableau is not standard, then it is not
straight. If T is not standard, it has an inversion, say aik is not <− ajk with
i < j. Let us suppose that this inversion is the left-most one that exists in the
tableau. If aik has no element immediately to the left of it, then there is no
way the inversion is unflippable, so T is not straight. If we do have an element
immediately to the left of aik , say bik−1 , then bik−1 lies above an element cjk−1 ,
176 Weyl and Schur modules
since our shape is a skew-shape. But since the k th column is the first one in
which we have an inversion, the column to the left of it is <− -increasing. Hence
bik−1 <− cjk−1 <+ ajk (the last inequality holding due to the row-standardness
of T ). Thus the inversion is flippable, and T is not straight. 2
From now on, we will focus our attention on Weyl modules; the appropriate
changes for Schur modules will mostly be left to the reader. In the Weyl case,
<+ -increasing means weakly increasing, while <− -increasing means strictly
increasing.
such l, alk > ajk , hence we have an inversion. Since (by maximality of i), this
inversion must be unflippable, we run through the usual argument to conclude
that µj > µl . Again we conclude that j = n, and the fact that our shape is of
type ≥ n − i − 1.
Let us observe too that if we have i < l < l < j, then we must have alk < al k .
Because if not, then (by maximality of i), the inversion has to be unflippable,
and by the usual leftward descent argument, we would conclude that µl < µl ,
which can not be since l < j ≤ n.
So finally, let us assume that j − i = 1. In that case, we have no extra inform-
ation except that if aik−1 exists, then aik−1 ≤ ai+1k . Furthermore, we have no
conclusion to draw about the type of the shape. We will, for the sake of simpli-
city of writing, refer to this flippable inversion of our non-straight tableau as the
extreme flippable inversion of T .
There are a few more observations we should make about these extreme flip-
pable inversions. As we have seen, we have aik−1 ≤ ajk if aik−1 exists. Let α
be the minimum value such that ajα exists and either aiα−1 does not exist, or
aiα−1 ≤ ajα . If row i does not protrude further to the left than row j or, what
is the same, µj ≤ µi , then α = µj + 1. If, on the other hand, µj > µi (in which
case we must have j = n), then we see that µn + 1 ≤ α ≤ k. This number α will
come into play when we describe our straightening algorithm. We should note
that if ajα−1 exists, then ajα−1 = ajα since we have ajα−1 < aiα−2 ≤ aiα−1
≤ ajα .
Our second example above suggests that we ask what happens if aik−1 = aik .
We saw in that example that we scooped up all of the second row along with
the third row in order to avoid inconvenient coefficients when we diagonalized.
We have to take this situation into consideration; it is a rather special case, but
merits special attention.
If we have aik−1 = aik , then we have aik−1 = aik = ajk (since aik−1 ≤ ajk ).
If ajk−1 exists, then we have aik−1 ≥ ajk−1 , so aik−2 must exist and be greater
than ajk−1 . Thus, in this case, our α = k, and we also see that we must have
j = n. (If ajk−1 does not exist, we certainly have α = k and j = n.)
This analysis of the type of configuration that can yield a non-straight tableau
is what informs our straightening procedure. But before we actually describe this
procedure, and maintain that applying it actually makes our tableaux “better”,
we have to say what “better” means.
Just to be sure that there is no confusion about this definition, let us look at
x3 x3 x3
the tableau, x4 x4 x4 . The column word is uT = x4 x2 x4 x3 x4 x3 x3 ; the
x2
modified column word is wT = x4 x4 x3 x2 x4 x3 x3 , and the reversed column word
is wT = x4 x2 x3 x4 x3 x4 x4 .
Notice that the tableau above appeared in Example VI.7.5, but there it was
being treated as an element of Kλ/µ . Here, however, we are regarding the tableau
purely combinatorially.
In Examples VI.7.4 and VI.7.5 above, the tableaux that were produced via
straightening were all less than the original tableaux we started with.
We are now set to show that the straight tableaux generate our Weyl module
or, what is the same, that every tableau can be written as a linear combination
(with integer coefficients) of straight tableaux. This is what we mean by the
straightening procedure referred to earlier. We proceed as we have done before:
we take a tableau, T , assume that it is not straight, and then show that it is
a linear combination of tableaux, T , which are all smaller than T in the order
relation just defined. Since we are dealing with finite sets, this will do the job.
It will then remain to show that the straight tableaux are linearly independent,
and hence form a basis for the Weyl module. We will always assume that we
start with row-standard tableaux; when we read a tableau as an element of the
Weyl module, if it were not row-standard, making it so would not change the
element.
Let T be a (row-standard) non-straight tableau, and assume that its extreme
flippable inversion occurs in adjacent rows, i and i+1 < n, where n is the number
of rows in the tableau. Focusing our attention just on those two rows, we have
the situation:
t at+1 ··· ak−1 ak ··· aq
T = ,
b1 ··· bt+1 ··· bk−1 bk · · · bq
with the inversion ak ≥ bk . (We are taking liberties here, calling just the two
rows T , but we think the meaning is clear. We have also changed the names of
the elements in the interests of clarity.)
We have drawn the diagram as though the cell immediately to the left of ak
exists; of course that need not be the case. But with our assumptions, t ≥ 0 and
we cannot have ak−1 = ak . We now let w ∈ Dk−1−t be the word at+1 · · · ak−1
(meaning that if there are repeats they are considered as divided powers rather
than simple powers), we let v ∈ Dq+1+ be the word b1 · · · bk · · · bk+ ak · · · aq ,
where bk = · · · = bk+ = bk++1 ( may be zero), and u ∈ Dq −k− the word
consisting of bk++1 · · · bq .
180 Weyl and Schur modules
We now have to be a bit more careful about our definition of the word w: let w
be the word consisting of the elements in the top row to the left of ak−δ , where δ is
such that ak−δ−1 < ak−δ = · · · = ak . We let v be the word bα · · · bk+ ak−δ · · · aq ,
that is, we do not include the terms bs+1 · · · bα−1 . Finally, we let u be the
word bs+1 · · · bα−1 bk++1 · · · bq . We have w ⊗ v ⊗ u ∈ Dk−δ−1 ⊗ Dq−α+2+δ+ ⊗
(k−δ−1) (k++1−α)
Dq −s+α−(k++1) . Now let us consider ∂2,1 ∂3,2 (w ⊗ v ⊗ u). Again
using the Capelli identities, we see that
(k−δ−1) (k++1−α)
k+−α
(k++1−α−h) (k−δ−1−h) (h)
∂2,1 ∂3,2 = (−1)h ∂3,2 ∂2,1 ∂3,1
h=0
(α−−2−δ) (k−α++1)
+ (−1)k+−α+1 ∂2,1 ∂3,1 .
The important thing to notice here is that the terms under the summation
sign are all in the kernel of the Weyl map (since they all involve positive powers
of ∂32 ), so that under the Weyl map, the sum
(k−1−δ) (k++1−α) (α−−2−δ) (k−α++1)
∂2,1 ∂3,2 − (−1)k−α++1 ∂2,1 ∂3,1
applied to w ⊗ v ⊗ u gives zero. If one analyzes the tableaux that show up under
this sum, one sees that in addition to our tableau, T , one gets tableaux with w
in the top row and the terms of v spread between the top and bottom rows (as
in the previous situation), and other tableaux in which the word w gets spread
around the top and bottom rows. Because of the inequalities we observed at
the beginning of this discussion, we see that all the tableaux that appear, other
than T , have column words greater in the lexicographic ordering than that of T .
Hence, we have again “straightened” our original tableau, and we have shown
that the straight tableaux generate the Weyl module for an almost skew-shape.
Proof This is the same as saying that in each column, there are no repeats.
Suppose for some i < j we had aik = ajk . Then straightness implies that aik−1 >
ajk and this is clearly impossible. 2
Proof Certainly this is the case if l = γj . If l = γj , and if the assertion were not
true, then some xjs , with s > l, must be in that box. But xjs > xjl , so we have
an inversion. However, there is either no element to the left of the box occupied
by xjs , or there is and its value is less than or equal to xjl . Therefore, this is a
flippable inversion, contradicting the straightness of our tableau. 2
We stress again that all of what we have said makes sense if we think of the
tableau simply as a filled-in diagram. Whether the filled-in diagram is to be
regarded as an element of the Weyl module or not, depends upon the context in
which we are using it.
What we propose to do now is start with a basis element of Λγ1 ⊗· · ·⊗Λγm , and
associate to it, when possible, a filling of the diagram which is straight and whose
reverse column word is the given basis element we started with. When B. Taylor
introduced this algorithm[83], he called it the straight-filling algorithm, and
the resulting filled diagram a straight filling. In fact, we will see that if this
procedure does not produce a straight filling, then there is no straight tableau
with that given reverse column word.
Algorithm (Taylor Algorithm)
Let D be the diagram of an almost skew-shape with columns c1 , . . . , cm , and let
w = x11 ∧ · · · ∧ x1γ1 ⊗ · · · ⊗ xm1 ∧ · · · ∧ xmγm be a basis element of Λγ1 ⊗ · · · ⊗ Λγm .
Arrange x11 , . . . , x1γ1 in increasing order in column c1 . Next, place x21 in the
first box of c2 which either has no neighbor to its immediate left, or has such
a neighbor whose entry is less than or equal to x21 . If there is no such box in
c2 , the output of the algorithm is “no straight filling”. If there is such a box, fill
it with x21 . Assuming that we have placed x21 , . . . , x2j , we place x2j+1 in the
first empty box of c2 which either has no neighbor to its immediate left, or has
such a neighbor whose entry is less than or equal to x2j+1 . Again, if there is no
such box, our output is “no straight filling”; if there is, we fill it with x2j+1 . We
continue in this way with the remaining columns, obtaining an output of “no
straight filling”, or a filling which we shall call T (w).
Let us look at an example or two. As our shape, we will take one we have used
earlier, namely (4, 3, 2)/(1, 0, 1). This has three rows and four columns, with
γ1 = 1, γ2 = 3, γ3 = 2, γ4 = 1. The word w1 = x4 ⊗ x1 ∧ x2 ∧ x4 ⊗ x4 ∧ x5 ⊗ x6
Straightening and the straight basis theorem 183
x1 x4 x6
produces the filling x4 x4 x5 , while the word w2 = x4 ⊗ x1 ∧ x2 ∧ x5 ⊗
x2
x1 x4 x4
x4 ∧ x6 ⊗ x4 produces the filling x4 x5 x6 . On the other hand, words
x2
such as x4 ⊗ x1 ∧ x2 ∧ x4 ⊗ x5 ∧ x6 ⊗ x4 or x5 ⊗ x1 ∧ x2 ∧ x4 ⊗ x4 ∧ x6 ⊗ x4 produce
“no straight filling.”
We point out a few things about this algorithm. First of all, if we start with a
straight tableau, T , with reverse column word w (written as a basis element of
our tensor product of exterior powers), then T (w ) will clearly have the reverse
column word w , if T (w ) exists. But from our second fact above, T (w ) clearly
does exist and equals T . We therefore see that if two straight tableaux have the
same reverse column word (and hence the same modified column word), then
they are equal. Hence the straight tableaux are precisely those whose reverse
column words produce a successful outcome of the straight-filling algorithm.
where the map labeled “inc” is the inclusion map. In addition to this inclusion
map, we also have the projection map proj:Λγ1 ⊗ · · · ⊗ Λγm → Λ(λ/µ). If we can
show that the composite map:
inc ωλ/µ proj
D(λ/µ) −→ Dp1 ⊗ · · · ⊗ Dpn −→ Λγ1 ⊗ · · · ⊗ Λγm −→ Λ(λ/µ)
Since D(λ/µ) and Λ(λ/µ) are both free modules with selected bases, we will
show that the composite map above is an isomorphism by showing that the
matrix of this map, with respect to a suitable total order of the bases, is upper
triangular with ones on the diagonal. However, to make the picture of what we
are about to do clear, we first give an example.
We will take our favorite almost skew-shape: λ/µ = (4, 3, 2)/(1, 0, 1) whose
(2)
diagram is D = . Consider the element x1 x4 x6 ⊗ x4 x5 ⊗ x2 ∈ D3 ⊗
x1 x4 x6
D3 ⊗ D1 ; under our Weyl map it goes to the tableau T = x4 x4 x5 ,
x2
which is straight, so this element is a basis element of D(λ/µ). For convenience of
writing, let us agree to replace xi by its index, i. Also, instead of writing wedges
for the product in the exterior algebra, let us simply juxtapose, and instead
of writing the tensor product symbol, we will simply write a slash. Using this
shorthand, we see that in Λ1 ⊗ Λ3 ⊗ Λ2 ⊗ Λ1 , T is the sum of elements:
4/142/45/6 + 4/142/65/4 + 4/152/64/4 + 4/452/14/6 + 4/452/64/1
+ 5/142/64/4 + 4/642/15/4 + 4/642/45/1 + 4/652/14/4 + 5/642/14/4.
Of course, in calculating out all the terms that occur as summands of T , many
are zero, as any repeat in a column is automatically zero in the exterior powers.
What we are left with, then, is a sum of terms which, but for a sign, are basis
elements of Λ1 ⊗ Λ3 ⊗ Λ2 ⊗ Λ1 . If we rewrite these terms so that the sequences
between slashes are increasing, we can ask which of them yields a straight filling
under the Taylor algorithm. With very little difficulty one sees that there are
only two: the “original” filling, and the one that corresponds to 4/125/46/4,
1 4 4
whose associated filling is 4 5 6 .
2
We see, then, that the composite map applied to our element yields
4/142/45/6 + 4/152/64/4.
Let us notice that the second filling was constructed from the term we denoted
4/152/64/4, and the reversal of order in the third column was due to the Taylor
algorithm. If we write T and T for the fillings corresponding to the two elements
above, and uT , uT , their corresponding column words, we see immediately that
uT < uT in the lexicographic order; thus T > T . If, now, we let T be the
straight filling of w (T ) (which is pictured immediately above), then uT < uT ,
so T > T .
This suggests the order we should put on the basis elements of Λ(λ/µ) and
D(λ/µ). Namely, we order the basis elements of Λ(λ/µ) according to the lexico-
graphic order on the column words of their straight fillings, and we order the
Straightening and the straight basis theorem 185
Proposition VI.7.8 The composite map proj ωλ/µ inc : D(λ/µ) → Λ(λ/µ) is
an isomorphism.
Proof The main argument in this proof is to show that with the ordering of the
basis elements that we have defined, the matrix of the map is upper triangular
with ones on the diagonal.
To this end, consider a basis element X = X1 ⊗ · · · ⊗ Xn ∈ D(λ/µ), and use
it to fill the diagram of λ/µ. Its image under the Weyl map is then the tableau
that we get using that filling (see example above). This tableau is straight, since
the hypothesis that X is in D(λ/µ) tells us precisely that, and the value of this
tableau in Λγ1 ⊗· · ·⊗Λγm is a sum of elements which are obtained by first diagon-
alizing each Xi into Λ1 ⊗ · · · Λ1 , and then multiplying the entries in each column.
% &# $
pi
To clarify this procedure, visualize the diagram as filled with the rearranged
terms; thus we obtain rows whose entries are permutations of the original row,
and the “columns” to which we refer are the columns of the diagram. (Again,
we recommend that you look at the example above). Because the multiplication
is occurring in an exterior algebra, any repeats will yield zero, so the summands
that remain will be those in which no repeats appear in any of the columns.
There are two facts that are fairly clear: the original filling is among the sum-
mands (the identity permutation on the rows is among the permutations of the
rows that we are considering); the column word of any of the fillings that occur,
other than the original one, is greater than that of the original one. This last can
be proven easily by observing that if there is any change in the first column, it
must be due to the substitution of one of the original entries by a larger one in
its row. A similar argument, column by column, proves the assertion.
The next thing we must do is apply the projection map to the surviving
summands; this now gives us the sum of those summands whose reverse column
word yields a straight filling. But it is easy to see that if T is any filling whose
reverse column word yields the straight filling, T , then the column word of T is
greater than that of T in the lexicographic order. For any change in any column
will be the result of moving a larger element in a column below a smaller one in
that same column (and one can now argue column by column as above).
As a result we see that the matrix of this map is indeed upper triangular, with
ones on the diagonal, and the proposition is proven. 2
We now have the straight basis theorem, as well as the presentation theorem
for almost skew-shapes.
1. The Weyl module, Kλ/µ (F ), is a free R-module with basis consisting of the
straight tableaux in a basis of F .
2. The map θλ/µ : K̄λ/µ (F ) → Kλ/µ (F ) is an isomorphism.
3. The functor Kλ/µ (F ), considered as a functor of F , is universally free.
Proof Our proposition above tells us that the straight tableaux are linearly
independent; in fact they are in one-one correspondence with the basis elements
of D(λ/µ) and of Λ(λ/µ). Since we already know that they generate Kλ/µ , they
form a basis of the Weyl module. This disposes of 1.
The map θλ/µ is clearly a surjection, so that the cosets of the basis elements
of D(λ/µ) generate K̄λ/µ . But it is clear that these elements are also linearly
independent, so they form a basis of K̄λ/µ . Hence the map θλ/µ must be an
isomorphism. This takes care of 2.
Finally, since K̄λ/µ is the cokernel of a natural map between two univers-
ally free functors, it must be universally free. But from 2, we conclude that
Kλ/µ is also universally free. (For a reminder of what universal freeness is, see
Subsection III.1.2.) 2
• In a row-standard tableau, two elements aik , ajk (with i < j) in the same
column are said to form (or be) an inversion if they violate column-standardness.
That is, if aik > ajk . The inversion is said to be unflippable if there is an element
in the tableau, bik−1 , immediately to the left of aik , such that bik−1 ≥ ajk .
Otherwise the inversion is called flippable. The row-standard tableau is said to
be straight if every inversion is unflippable.
In a row-standard tableau of negative letters, the straightening is a bit easier
than is the case for positive letters, as there can be no repeated elements in
any given row. Hence, whereas in the straightening algorithm for positive letters
we had to scoop up elements in a row that repeated, here our straightening
procedure will not require that. Except for that, our straightening of tableaux
goes through as before, and we have the fact that the straight tableaux generate
the Schur module. To prove linear independence of the straight tableaux, we
must modify the Taylor algorithm.
Since the Schur map goes from tensor products of exterior powers to tensor
products of symmetric powers, and a basis of the tensor product of symmetric
Weyl–Schur complexes 187
VI.7.6 Duality
As anticipated in Remark III.1.5 for the special case of hooks, there is a duality
between some Weyl and Schur modules. It is well known, [3], that if A is the
shape matrix of a skew-shape, λ/µ, and A is its transpose (again the shape matrix
∗ ∼
of a skew-shape), then KA (F ) = (LA(F ))∗ . The proof depends in part on the
fact that for such shapes, the modules in question are universally free. Now, if
A is the shape matrix of an almost skew-shape of positive type, its transpose
is no longer of the same kind. Therefore if we want an isomorphism like the
one stated, we would at least have to saturate the class of almost skew-shapes
with respect to transposition, and develop all of the preceding material for that
larger class of shapes. Since all of the shapes in that class would be row-convex,
and the straightening techniques and algorithms used in this chapter apply to
row-convex shapes, one could probably arrive at such a duality statement.
while
(∗∗) (D(ϕ1 ) ⊗ D(ϕ2 ))k = Dα (ϕ1 ) ⊗ Dk−α (ϕ2 )
α
= Dβ (G1 ) ⊗ Λα−β (F1 ) ⊗ Dγ (G2 ) ⊗ Λk−α−γ (F2 ).
α,β,γ
If we set γ = α − β, we see that the last sum in (∗∗) rearranges into the last sum
in (∗). 2
This clearly allows us to think of a basis of Rn as, say, positive places, and
to develop the letter-place context for this situation. In fact, we will think of a
basis of F , say F, as an ordered set of negative letters, and a basis of G, which
we will write G, as an ordered set of positive letters. For convenience (it is easy
to show that our choice makes no essential difference to our discussion), we shall
assume that every element of G precedes all the elements of F, and so we have
a total order on G " F.
We will use the customary double tableau notation, but since the “letters”
are basis elements of D(G) ⊗ Λ(F ), we should explain what we mean by an
expression such as (w ⊗ u|1(k1 ) 2(k2 ) · · · n(kn ) ), where w ∈ Dk (G), u ∈ Λl (F ) and
k + l = k1 + · · · + kn . To be consistent with our previous notation for letter-place
identities, we set
(w ⊗ u|1(k1 ) 2(k2 ) · · · n(kn ) ) = (w(s1 ) ⊗ u(t1 )|1(k1 ) ) · · · (w(sn ) ⊗ u(tn )|n(kn ) ),
where the sum ranges over all pairs si , ti such that si + ti = ki . We have written
w(si ), u(ti ) to indicate the components of the n-fold diagonalization of w and u
of degrees si , ti . Again, an example will serve to make this much clearer.
Once we have this notation at our disposal, we are set to use our old double
tableau notation. If A stands for the multi-signed alphabet G " F, the notation
and definitions of Subsection VI.7.1 apply to our situation.
At this point, it must be fairly clear to the reader that the same “straightening”
arguments we have used before will work here to show that the double standard
tableaux generate the n-fold tensor product of D(ϕ). We refer the reader to
Section 4 of Appendix A to see how to prove linear independence in much the
same way as before. An illustrative example will also be found there. As a result
we have the theorem:
Theorem VI.8.4 Let ϕ : G → F be a morphism of free R-modules. Then the
double standard tableaux discussed above form a basis for D(ϕ ⊗ Rn ). The same
result obtains for Λ(ϕ ⊗ Rm ).
Clearly, one can now proceed with place polarizations as in Section VI.4.
Because the polarizations operate only on the places, all the structure attached to
the letters is preserved by them. In particular, the complex structure that we have
imposed on our algebra is preserved. As a result, given an almost skew-shape,
λ/µ, (or any shape for that matter), we can define the map
∆λ/µ (ϕ) : Dp1 (ϕ) ⊗ · · · ⊗ Dpn (ϕ) → Λγ1 (ϕ) ⊗ · · · ⊗ Λγm (ϕ)
as we did the Weyl and Schur maps, polarizing the positive places (basis of Rn )
to the negative places (basis of Rm ) according to the recipe dictated by the
diagram of λ/µ. (The pi and γj are read from the diagram of λ/µ in the usual
way.) This, by our observation above, is a map of complexes, which we call the
Weyl–Schur map (or W–S map); as a result, its image is also a complex.
Definition VI.8.5 The image of the map ∆λ/µ is the Weyl–Schur (W–S)
complex, Kλ/µ (ϕ), associated to the map ϕ and the shape λ/µ.
Remark VI.8.6 It is important to point out here that, since the map ∆λ/µ
does not depend on the map ϕ at all, the chains of the complex Kλ/µ (ϕ) depend
only on the shape λ/µ and the modules G and F .
With this definition, we can proceed to define the complex Rel(λ/µ) as we did
for Weyl and Schur modules, as well as the map of complexes, λ/µ , whose image
is in the kernel of ∆λ/µ (ϕ). We denote again, by K̄λ/µ (ϕ), the cokernel of the
map λ/µ and by θλ/µ the naturally induced map of complexes from K̄λ/µ (ϕ)
to Kλ/µ (ϕ).
Given our signed inequalities, namely, <+ and <− , our definitions of standard
and straight apply to the tableaux of shape λ/µ, filled with the elements of G "F,
and our analysis of flippable and unflippable inversions goes through without
change. Thus the straightening procedure we used to show that the straight
tableaux generate goes through using <+ and <− on the multi-signed alphabet
G " F. Similarly, the Taylor algorithm goes through without change as long as
we use the inequalities above. In fact, one can see that the “modifications” made
Weyl–Schur complexes 191
in discussing the Taylor algorithm for Schur modules rather than Weyl modules
were really unnecessary if we had used the <+ and <− inequalities throughout.
As a result, we can state the following theorem without further proof:
Theorem VI.8.7 Let ϕ : G → F be a morphism of free R-modules, and λ/µ
an almost skew-shape. The following statements are true:
1. The Weyl–Schur complex, Kλ/µ (ϕ), is a free R-complex with basis consist-
ing of the straight tableaux in a basis of F union a basis of G.
2. The map θλ/µ : K̄ λ/µ (ϕ) → Kλ/µ (ϕ) is an isomorphism.
This page intentionally left blank
VII
SOME APPLICATIONS OF WEYL AND SCHUR MODULES
In the previous chapter we developed the basic definitions of Weyl and Schur
modules, as well as Weyl–Schur complexes. In this chapter, we will give
indications of how we may apply some of these ideas.
In the first two sections, we deal with very basic material, and go into
some detail, as the results are fundamental for most of the applications of
characteristic-free representation theory that we will see. The remainder of the
chapter takes a more cavalier approach to the results considered; we include a few
historical notes and rely even more heavily than heretofore on recent reference
material.
What we intend to prove is that the kernel of this surjection is the W-S complex,
Kλ /µ (ϕ), where
λ /µ = (p1 , . . . , pi−1 , pi + ti − s + 1, pi+1 , . . . , pn−1 , pn − ti + s − 1;
t1 , . . . , ti−2 , ti−1 + ti + s − 1, s − 1, ti+1 , . . . , tn−2 , −(tn−2 + · · · + ti )).
While the above description of λ /µ is very graphic, it is also very unwieldy.
Also, our insistence that in the case of a skew-shape we have µn = 0, but for
an almost skew-shape we have µn−1 = 0, makes for a certain amount of distinct
case analysis in our discussions. To avoid this, we will simply start with our
almost skew-shape, λ/µ, define our shape λ /µ to be (λ1 + 1, . . . , λn−1 + 1, λn )/
(µ1 + 1, . . . , µn−1 + 1, µn ), and then our shape λ /µ is simply:
(λ1 + 1, . . . , λn−1 + 1, λn )/(µ1 + 1, . . . , µi−1 + 1, µn , µi+1 + 1, . . . , µn−1 + 1, µi + 1).
same, say W. It is also clear that the domain of ∆λ /µ is N, where N is as
defined above. Therefore, if we can prove that the diagram:
(t −s+1)
∂n,ii
N / M
one can move much more to the left (still by only one step). In fact, a similar,
but much easier argument than that used to prove Theorem VII.1.2 yields the
following result.
where
and
Discussion. We will only discuss the considerations that lead to the definition
of the filtration.
First, we order the basis for G ⊕ F , with G = G1 ⊕ G2 and F = F1 ⊕ F2 ,
by making the basis elements of Gi precede those of Fi , and the basis elements
of G1 ⊕ F1 precede those of G2 ⊕ F2 . For notational convenience, let us denote
the basis of G1 ⊕ F1 by U, and the basis of G2 ⊕ F2 by V. Since for a skew-
shape, “straight” tableaux are the same as “standard” tableaux, we see that
row-standardness of a tableau for Kλ/µ (ϕ) places the elements of U before those
of V, while column-standardness requires that the “separators” between the first
198 Some applications of Weyl and Schur modules
and the second set of elements in each row together form a partition ν that
contains µ and is contained in λ. We may illustrate this situation as follows:
u v
(S) u v .
u v
This leads us to define the filtration on Kλ/µ (ϕ) as follows.
For each partition ν with µ ⊆ ν ⊆ λ, define
Mν (Dλ/µ (ϕ)) = im Dσ/µ (ϕ1 ) ⊗R Dλ/σ (ϕ2 ) → Dλ/µ (ϕ)
ν⊆σ⊆λ
Ṁν (Dλ/µ (ϕ)) = im Dσ/µ (ϕ1 ) ⊗R Dλ/σ (ϕ2 ) → Dλ/µ (ϕ)
ν⊂σ⊆λ
Mν (Kλ/µ (ϕ)) = ∆λ/µ (ϕ) Mν (Dλ/µ (ϕ))
Ṁν (Kλ/µ (ϕ)) = ∆λ/µ (ϕ) Ṁν (Dλ/µ (ϕ)) ,
where the top two maps are essentially the inclusion maps, and ∆λ/µ (ϕ) is the
usual map. The proof that the above is our desired filtration and that
Mν (Kλ/µ (ϕ)) / Ṁν (Kλ/µ (ϕ)) ∼
= Kν/µ (ϕ1 ) ⊗R Kλ/ν (ϕ2 )
is actually an isomorphism is found in Reference [3]; the reader should be warned
that there the bold-faced D is instead written as Λ.
Let us now look at what the situation would be if instead of a skew-shape we
had started with an almost skew-shape (still denoted by λ/µ). Keeping our same
ordering of the basis, and looking at a straight tableau, we see that as in the
previous case, our row-standardness has all the basis elements of U preceding
those of V. An analysis of straightness in the columns, however, tells us that the
line of demarcation in each row gives us, until we reach the penultimate row,
what looks like a subpartition of λ which contains µ. But straightness allows the
possibility that in the last row, we have some basis elements of U lying under
some elements of V in a higher row, provided that this higher row protrudes
beyond the bottom row. As a result, these demarcations delineate an almost
partition, rather than a partition. Again, it is useful to illustrate this situation
with the following diagram:
u v
u v
(AS) .
u v
u v
One approach to dealing with this could be to consider a larger class of shapes
than the almost skew-shapes. For instance, we could consider shapes λ/µ, with
Resolution of determinantal ideals 199
or maximal, minors of the matrix. But if, using the notation of Chapter III, we
have the map f : A → B, with A = Rm , B = Rn , then we can see that the
cokernel of the map mq (f ) : Λq A ⊗R Λr B → Λq+r B has support equal to that of
the ideal generated by the minors of order n − r (where mq (f ) is the composition
Λq f ⊗R id
Λq A ⊗R Λr B −→ Λq B ⊗R Λr B → Λq+r B). Thus there was the hope that
one could find some way of resolving this cokernel for all r in the same way as
one did the case when r = 0, that is, the case of the maximal order minors. Alas,
the problem turned out to be far more complex than imagined, but fortunately
it did lead to a number of interesting developments, some of which we will try
to cover in the subsequent subsections.
Sν ⊗R F → Sν ⊗R S1 ⊗R G∗ → Sν+1 ⊗R G∗ ,
where the last map is just the multiplication in the symmetric algebra.
But the map F to F ⊗R G ⊗R G∗ simply sends an element x ∈ F to the
element x ⊗ cG ∈ F ⊗R G ⊗R G∗ , where cG ∈ G ⊗R G∗ is the element (as in
Chapter V) corresponding to the identity map in HomR (G, G) under the natural
identification of G⊗R G∗ with HomR (G, G). If we now choose bases {x1 , . . . , xm }
for F , {y1 , . . . , yn } for G, and set Xij = xi ⊗ yj ∈ S1 , it is easy to see that the
matrix corresponding to our map, using {1 ⊗ xi } and {1 ⊗ yj } as bases for F and
G∗ , is simply (Xij ). The upshot of all of this is that the “generic” map is defined
without recourse to a basis, so that it seemed natural to suppose that the modules
that would occur in a resolution of the ideal, Ip , generated by minors of order p
(which also can be defined without recourse to basis), should be independent of
a choice of basis, and hence should be representations of GL(F ) × GL(G).
Resolution of determinantal ideals 201
However, it was not until the mid-1970s that A. Lascoux found a way to expli-
citly use representation theory to describe the resolutions of these determinantal
ideals [64]. Working in characteristic zero, where every representation of the gen-
eral linear group splits into a direct sum of irreducible representations, he was
able to write down the GL(F ) × GL(G)-irreducible components of the repres-
entation modules Xkp that occur as the chains in dimension k of the resolutions
of the determinantal ideal Ip , the ideal of p × p minors of the generic m × n
matrix. Since the irreducible representations of GL(F ) × GL(G) are of the form
M ⊗R N , where M and N are irreducible representations of GL(F ) and GL(G),
resp., and the irreducible representations would be of the form Lλ F and Lµ G
for partitions λ, µ, the question to be answered is: for which pairs of partitions
λ, µ do the tensor products Lλ F ⊗R Lµ G appear as direct summands in Xkp ?
A. Lascoux determined what this family of pairs is by using a clever geometric
argument. First he “desingularized” the variety of Ip , by going up to the appropri-
ate Grassmannian; the ideal of this desingularized variety has a Koszul complex
as its resolution. Then, by Bott’s theorem on the vanishing of the cohomology of
homogeneous vector bundles ([14], theorem 4), the higher direct image functors
on the Koszul complex can be computed, and the resulting complex is the min-
imal resolution of Ip . While the desingularization technique holds in arbitrary
characteristic, Bott’s theorem is valid only in characteristic zero, which partially
explains why the Lascoux approach seems to be limited to that case. (We will
return to this point in a later section.) This geometric approach was originated
by G. Kempf in the 1970s [57]; a full and very clear discussion of this and similar
techniques can be found in J. Weyman’s book [86].
It should be pointed out here, for the sake of historical accuracy, that an
explicit description of the boundary maps of the Lascoux resolutions was not
given by A. Lascoux. Subsequent work by H. Nielsen [68] and P. Pragacz and
J. Weyman [70] was needed to clear this up.
To describe the partitions that do appear in the Lascoux resolutions, we
need some additional notation and terminology. We say that a partition, λ, has
Durfee square of size k if λi ≥ k for i = 1, . . . , k, and λk+1 ≤ k. We denote
the transpose partition of λ as usual (see Chapter VI), by λ, so that we write
= (λ
λ 1 , . . . , λ
t ) where λ
j = #{i| λi ≥ j}. It is clear that λ and λ
have the same
Durfee square size. Finally, if λ and µ are partitions, we denote by λ + µ the
partition (λ1 + µ1 , . . . , λq + µq ).
We are now in a position to state (without proof) what the terms of the
Lascoux resolution are.
Theorem VII.3.1 Let Xkp denote the module of k-chains in the resolution of
Ip . Then
Xkp = Lλ+(p−1)k(λ) F ⊗R Lλ+(p−1)
k(λ) G,
λ∈Λk
202 Some applications of Weyl and Schur modules
where k(λ) is the Durfee square size of λ, (a)k(λ) denotes the rectangular partition
(a, . . . , a), and Λk is the set of partitions, λ, such that |λ| = k.
% &# $
k(λ)
of the terms that appear in characteristic zero (the term Z-form will be explained
in the next subsection) in order to push the construction through.
As in the previous discussion, we will skip the details, but two observations
emerged from this project: one was that the boundary map which was explicitly
constructed was not “natural,” that is it was not equivariant with respect to
the action of the general linear groups involved; another was that it was nev-
ertheless possible to build the terms of the resolution in a way suggested by
the Lascoux approach (although many more terms might have to be involved—
more about this in a later section). An important additional observation was
that fairly delicate arithmetic considerations were involved in the construction
of these complexes, and this led to an independent study of the Z-forms which
we will discuss in the next subsection.
VII.3.3 Z-forms
Let F be a free abelian group of rank m, and let F be its extension to the
rationals, that is, F = Q ⊗Z F . We know that D2 F and S2 F are both GL(F )-
representations , where GL(F ) means the general linear group over the integers,
Z. Furthermore, D2 F = Q ⊗Z D2 F, S2 F = Q ⊗Z S2 F , and these are GL(F )-
representations. We know that there is a GL(F )-equivariant map from D2 F to
S2 F , namely the composition:
∆ m
D2 F −→ F ⊗Z F −→ S2 F,
where ∆ is the diagonal map, and m is the usual multiplication map. However,
this is not an isomorphism of the two integral representations. Nevertheless, the
corresponding map over the rationals is an isomorphism. We therefore say that
D2 F and S2 F are Z-forms of the same rational representation (in this case, S2 F ).
So, we are led to make the following definition.
are exact, where the boundary map is given by diagonalizing the exterior powers
and multiplying the symmetric powers. But what happens if we replace the
204 Some applications of Weyl and Schur modules
where we still diagonalize the exterior powers and multiply, this time, into the
divided powers (something that we have avoided doing so far throughout this
book)? As the reader may strongly suspect, this complex is no longer exact;
the surprising thing, however, is that, counting from the left, it is exact up to
the middle of the complex, that is, from t = 0 to t = [(l − 1)/2] where [x]
indicates the integral part of x ([4], proposition 2.22). As a result, the cycles
of this complex are Z-forms of the corresponding cycles of the complex above,
involving the symmetric powers, and these, as indicated in Chapter III, are just
the hooks.
Another, simpler, way to construct non-isomorphic Z-forms is the following:
Consider the short exact sequence
where K(k+1,1) is the Weyl module associated to the hook partition (k + 1, 1).
(We are leaving out the module F , as that is understood throughout.)
If we take an integer, t, and multiply Dk+2 by t, we get an induced exact
sequence and a commutative diagram:
0 → Dk+2 → Dk+1 ⊗Z D1 → K(k+1,1) → 0
↓t ↓ ↓
0 → Dk+2 → E(t; k + 1, 1) → K(k+1,1) → 0,
where E(t; k + 1, 1) stands for the cofiber product of Dk+2 and Dk+1 ⊗Z D1 .
Each of these modules is a Z-form of Dk+1 ⊗Z D1 , but for t1 and t2 , two
such are isomorphic if and only if t1 ≡ t2 mod k + 2 (see [2]). This says that
Ext1A (K(k+1,1) , Dk+2 ) ∼
= Z/(k + 2), where A is the Schur algebra of appropriate
degree (we will discuss the calculation of Ext in more detail in Subsection VII.4.1
on intertwining numbers).
Although we have not introduced the Schur algebra until now, it plays an
important role in representation theory as, classically, the Schur algebra of degree
k is the universal enveloping algebra for the homogeneous polynomial representa-
tions of degree d. For example, Dd and Λd are such representations of the general
linear group. As a result, when we speak of “equivariant maps” of representa-
tion modules, what we really mean are linear maps over the Schur algebra; the
Schur algebra is the underlying ring over which all of our representations are
modules.
where A denotes the appropriate Schur algebra over Z, and A denotes Z/(p)⊗Z A.
and, in Chapter III, we saw that the hooks were the cycles in this complex.
(We will soon see other ways of arriving at this fact.) In any event, to compute
Ext1A (K(k+1,1) , Dk+2 ) (this, after all, is the situation for λ = (k + 1, 1), d = 1
and µ = (k + 2, 0) of our discussion), we simply look at
From our discussion of weight submodules, we see that both of these abelian
groups are free of rank one; the question is to identify bases for these groups and
make the map explicit in terms of these bases. A simple calculation shows that
the basis for HomA (Dk+2 , Dk+2 ) is the identity map, α, say, while the basis for
HomA (Dk+1 ⊗Z D1 , Dk+2 ) is the multiplication map, say β. The map from Dk+2
to Dk+1 ⊗Z D1 , call it δ, is just the diagonal map. Thus, the map β is carried
(k+2) (k+2)
to βδ, and we see that βδ(x1 ) = (k + 2)x1 = (k + 2)α. The matrix
in question, then, is the one-by-one matrix, (k + 2), and so we get the result
announced earlier, namely that Ext1A (K(k+1,1) , Dk+2 ) ∼ = Z/(k + 2).
While this calculation was quite simple, this is not the case in general. As of the
current writing, the only complete results that are known are for two- and three-
rowed partitions, and only for Ext1A . Again we refer the reader to References [2],
[26], and [30] for explicit calculations leading to the known results, but as we
see, there is a strong incentive to get our hands on an explicit description of
projective resolutions of Weyl modules. The reader may be interested to know
that a newer approach to calculating the invariant factors of the matrices that
we do encounter involves the use of letter-place methods, but this development
is still in its infancy (see [22], section 4).
powers are the symmetric powers, and Dk+1 ⊗Z D1 splits up into the direct sum
of K(k+1,1) and Dk+2 . Over the integers, the exact sequence (†) not only does
not split, but that short exact sequence turns out to be the generator of Ext.
Clearly there is a great deal of arithmetic involved in our integral representations
of the general linear group over Z.
The following example will also illustrate how these Z-forms are directly con-
nected with resolutions of Weyl modules. In Section VII.6 we will see that a
projective resolution of K(k,2) is given by
Dk+1 ⊗Z D1
0 → Dk+2 → ⊕ → Dk ⊗Z D2 → 0,
Dk+2
where the map from each component of the split term is the appropriate diag-
onalization (possibly followed by multiplication), the map Dk+2 → Dk+1 ⊗Z D1
is the diagonal, and the map Dk+2 → Dk+2 is multiplication by 2. Thus the
cokernel of the left-hand map is the module we denoted by E(2; k + 1, 1) in
Subsection VII.3.3. We therefore have the short exact sequence
0 → E(2; k + 1, 1) → Dk ⊗Z D2 → K(k,2) → 0,
and over the rationals, the sequence would look the same, except that there we
could replace E(2; k +1, 1) by Dk+1 ⊗Z D1 , which gives the anticipated resolution
of the Weyl module in characteristic zero (see [64]). From the point of view of the
Grothendieck ring of representations, this tells us that while over the rationals
the class of K(k,2) may be represented as [Dk ⊗Z D2 ] − [Dk+1 ⊗Z D1 ], over the
integers it would be represented by [Dk ⊗Z D2 ] − [E(2; k + 1, 1)].
The thrust of the above discussions is that there are many reasons for trying
to find an explicit description of projective resolutions of Weyl modules.
these ideals were known to be perfect (i.e., their homological dimension is equal
to their depth) in every characteristic (see [38], [39]), the implications about
Betti numbers seemed perfectly (sic) natural. That is, while certain construc-
tions were universal, they lacked universal minimal free resolutions, but those
that were known “violators” were not perfect in all characteristics (such as the
projective plane, which is perfect in every characteristic but two [51], [72]). Some
attempts were made to prove this apparently true fact about the Betti numbers
by less cumbersome techniques and, in 1987, K. Kurano [62], proved that the
first Betti numbers of Ip were independent of the characteristic.
But in 1988, M. Hashimoto announced (in a preprint which later became Ref-
erence [47]) the surprising fact that for the generic 5 × 5 matrix, the second Betti
number of I2 in characteristic three is one greater than it is in characteristic zero.
(His paper [47] actually proved that the ideal Iq of the (q + 3) × (q + 3) matrix
has its second Betti number in characteristic three larger than in character-
istic zero.) Immediately following M. Hashimoto’s announcement, J. Roberts and
J. Weyman [73] published another proof that had the same geometric flavor as
the Lascoux construction of his resolutions of determinantal ideals (see also [86]).
We will not go into their proof in detail, but a few remarks are worth discuss-
ing. The first observation made by Roberts and Weyman is that the geometric
part of the Lascoux approach is characteristic-free. That is, the desingularization
that is used over the Grassmannian, and the fact that the ideal of this desin-
gularization is resolved by a Koszul complex, are valid over any ground ring. It
is in the application of Bott’s theorem, that is, the second step of the Lascoux
program, that the procedure breaks down (as one might suspect, since the Bott
theorem is not true in positive characteristic). But what Roberts and Weyman
do is identify the higher direct images that come up in the Bott theorem with
the homology of certain canonical subcomplexes of the Weyl–Schur complexes.
The homology of these complexes is directly related to the Betti numbers of the
determinantal ideals, and can be computed by means of representation theory
(at least theoretically).
In the Hashimoto example, starting with such facts as that the representation
Λ3 is contained in the representation K(2,1) in characteristic three, it is possible
to see that there is a linear two-cycle that cannot be covered by a degree one map.
Hence the second Betti number goes up in characteristic three. This observation
makes sense in the light of the specific construction of the resolutions involved;
we will forbear to go into those details here.
It suffices to say that this example of M. Hashimoto took the steam out
of the project of looking for universal minimal free resolutions of determin-
antal ideals. It did, however, open up a few new areas of investigation, and
did use in an essential way the characteristic-free approach to Weyl modules
and their resolutions. Finally, the relationship between Bott’s theorem and the
homology of certain subcomplexes of Weyl–Schur complexes that is evident in
the Roberts–Weyman approach to M. Hashimoto’s example should be studied
carefully.
Resolutions of Weyl modules 211
p+t+1 t+1 p t p
0→ → → → 0,
q−t−1 q q
and an induction argument on the number of overlaps between the two rows,
namely, the integer q − t as read from the above diagram, it was possible to give
a description of the desired type of resolution [1, 28] . This will be described in
Subsection VII.6.2.
The three-rowed case presented, of course, quite another aspect, for as we
know, the kernel of the analogous surjective map is an almost skew-shape about
which nothing was known at the time. In particular, a presentation of such shapes
was not even guessed at. To be precise, the analogue of the above sequence is
212 Some applications of Weyl and Schur modules
the following:
t1 p t1 p
0→ q+t2 +1 → t2 +1 q
t2 +1 r−t2 −1 r
t1 p
→ t2 q → 0.
r
and while one can try to do an induction on the integer r − t1 − t2 , the type
of shape has changed on us, that is, it has become an almost skew-shape rather
than remain a skew-shape, as it so conveniently did in the two-row situation.
Hence one must also try to resolve these “new” shapes.
The approach to this in Reference [2] was to resort to a fairly complicated
spectral sequence to prove a slightly weaker form of Theorem VII.1.2.
Despite the encouragement provided by the proven existence of resolutions
of the type desired, and the fundamental exact sequences which theoretically
provide a mechanism for an induction proof as in the two-rowed case, the methods
at hand still made the calculations almost impossible to deal with. As a result,
it was not until the 1990s that this problem was looked at again from the point
of view of letter-place methods, where the Weyl map and the presentation map
were seen to be place polarizations, and the Capelli identities could play a role
in simplifying and systematizing these calculations. And then the introduction
by B. Taylor of straight tableaux made it possible to develop our theory as we
have in this book, and to prove the theorem on fundamental exact sequences in
more general form.
One other tool that was essential in describing the resolutions was the bar
complex (see [29]). As the reader can see, the strategy in building the resolutions
lies in assuming that one knows what the terms look like for shapes that are
“less complex,” with “complexity” being measured by the number of overlaps
between rows. Theorem VII.1.2 allows us to reduce this complexity by presenting
our given shape in terms of shapes that are less complex (as we did in the proof
of Theorem VII.1.3). Assuming we have resolutions of these less complex shapes,
we have a map between these resolutions (this is where the projectivity of the
resolutions comes in), and the mapping cone of this map is then a resolution
of the shape we started with. By describing the resolutions in terms of the bar
complex, a complex which behaves well with respect to forming mapping cones,
this procedure of putting together resolutions by forming mapping cones becomes
manageable. In the rest of this section we will describe some of these techniques
in more detail.
The main components of the construction of such complexes are the exterior
algebra over Z, Λ(S), on a set of free generators, S, called the separators, an
algebra, A, an A-module, M , and the free product of Λ(S) with A. The algebra
Λ(S) has a natural Z2 -grading: if m is a monomial in Λ(S), that is, a product of
generators, we set |m| = 0 if m is the product of an even number of generators,
and |m| = 1 if m is the product of an odd number of generators.
Definition VII.6.1 The free product of the algebra A and the algebra Λ(S)
will be called the bar algebra on the algebra A, with set of separators S,
and denoted by Bar(A; S).
Definition
VII.6.2 If T is a non-empty finite subset of S, the operator ∂T
= x∈T ∂x , is called the T -boundary operator on Bar(A; S).
214 Some applications of Weyl and Schur modules
Thus, the free bar module on M with a single separator gives rise to the
classical bar complex, as one sees by replacing the symbol x by the symbol “|”.
What we have called Bar(M, A; S, i) are simply the chains in dimension i.
t p
(A)
q
is the image of Dp ⊗R Dq under the Weyl map. The “box map” referred to at
the very beginning of Section VI.6, and denoted by λ/µ , was described there
as the sum of place polarizations,
(k)
∂2,1 : Dp+k ⊗R Dq−k → Dp ⊗R Dq .
k>t k>t
If we let Z2,1 stand for the generator of a divided power algebra in one free
(k)
generator, we see that Z2,1 acts on Dp+k ⊗R Dq−k and carries it to Dp ⊗R Dq .
Thus, we may take the (t+ )-graded strand of degree q of the normalized bar
complex of this algebra acting on Dp+k ⊗R Dq−k (where the degree of the
216 Some applications of Weyl and Schur modules
second factor determines the grading) to get a complex over the Weyl module:
(t+k ) (k ) (k )
··· → Z2,1 1 xZ2,12 x · · · xZ2,1l+1 x ⊗R (Dt+p+|k| ⊗R Dq−t−|k| ) →
ki >0
(t+k1 ) (k ) (k )
Z2,1 xZ2,12 x · · · xZ2,1l x ⊗R (Dt+p+|k| ⊗R Dq−t−|k| ) → · · ·
ki >0
(t+k)
→ Z2,1 x ⊗R (Dt+p+k ⊗R Dq−t−k ) → Dp ⊗R Dq → 0,
k>0
where the symbol “x” is our separator variable as described above, and |k| stands
for the sum of the indices ki . Here, the boundary operator is ∂x or, what is the
same thing, is obtained by polarizing the variable x to the element 1. This, then,
describes a left complex over the Weyl module in terms of bar complexes and
letter-place algebra. We also know from the fact that the Weyl module is the
cokernel of the box map, that the zero-dimensional homology of this complex is
the Weyl module itself.
Now the question is: how do we show that this complex is an exact left complex
over the Weyl module? In other words, that it is in fact a resolution. One way,
is to produce a splitting contracting homotopy, which is what we will do here.
Another way is to use our fundamental exact sequences and a mapping cone
argument; we refer the reader to Reference [2] for this approach.
Definition VII.6.9 With our complex given as above, define the homotopy as
follows:
(t+k)
s0 : Dp ⊗R Dq → Z2,1 x ⊗R Dt+p+k ⊗R Dq−t−k
k>0
4
w 44 1(p) 2(k)
sends the double standard tableau to zero if k ≤ t, and to
w 4 2(q−k)
4
(k) w 44 1(p+k)
Z2,1 x ⊗ if k > t. For higher dimensions (l > 0),
w 4 2(q−k)
(t+k ) (k ) (k )
sl : ki >0 Z2,1 1 xZ2,12 x · · · xZ2,1l x ⊗R Dt+p+|k| ⊗R Dq−t−|k| →
(t+k1 ) (k2 ) (kl+1 )
ki >0 Z2,1 xZ2,1 x · · · xZ2,1 x ⊗R Dt+p+|k| ⊗R Dq−t−|k|
4
(t+k1 ) (k2 ) (kl ) w 44 1(t+p+|k|) 2(m)
is defined by sending Z2,1 xZ2,1 x · · · xZ2,1 x ⊗ to
w 4 2(q−t−|k|−m)
4 (t+p+|k|+m)
(t+k ) (k ) (k ) (m) w 44 1
zero if m = 0, and to Z2,1 1 xZ2,12 x · · · xZ2,1l xZ2,1 x ⊗
w 4 2(q−t−|k|−m)
if m > 0.
where we assume that the number of triple overlaps is ≤ 1. This means that
p3 ≤ t1 + t2 + 1. The important thing to notice about this class of shapes is that
our fundamental exact sequence is very special in that it does not entail any
almost skew-shapes. Namely, the exact sequence we are interested in reduces to:
0 → p1 , p2 + t2 + 1, p3 − t2 − 1; t1 + t2 + 1, −(t2 + 1))
→ (p1 , p2 , p3 ; t1 , t2 + 1) → (p1 , p2 , p3 ; t1 , t2 ) → 0. (VII.6.1)
(The notation we are using above is that used in Section VI.6.) But, since the
third row of (p1 , p2 + t2 + 1, p3 − t2 − 1; t1 + t2 + 1, −(t2 + 1)) does not overlap
with the top row, we see that the Weyl module associated to this shape is iso-
morphic to that associated to the skew-shape (p1 , p2 + t2 + 1, p3 − t2 − 1; t1 +
t2 + 1, 0). Hence we have the exact sequence of skew-shapes:
0 → (p1 , p2 + t2 + 1, p3 − t2 − 1; t1 + t2 + 1, 0)
→ (p1 , p2 , p3 ; t1 , t2 + 1) → (p1 , p2 , p3 ; t1 , t2 ) → 0. (VII.6.2)
where the maps Dp1 +t1 +k ⊗R Dp2 −t1 −k ⊗R Dp3 → Dp1 ⊗R Dp2 ⊗R Dp3 are the
kth divided
power of the place polarization from place 1 to place 2, and the
maps l>0 Dp1 ⊗R Dp2 +t2 +l ⊗R Dp3 −t2 −l → Dp1 ⊗R Dp2 ⊗R Dp3 are the lth
divided power of the place polarization from place 2 to place 3. Again taking our
cue from the two-rowed case, but recognizing that we now need two separators
instead of one, we introduce two generators, Z2,1 and Z3,2 , with their divided
218 Some applications of Weyl and Schur modules
(k)
where the symbol ∂31 means the divided power of the usual place polarization,
and its presence on the right means that it is to be brought to operate on the
element of M that is to the right of the tensor product sign. For example, if
W = W ⊗ m, with m ∈ M, then ∂31 W means W ⊗ ∂31 (m). To illustrate: if
(k) (k)
(c)
we have W = Z21 x ⊗ m, then
(a) (b)
(b−k) (a−k) (c) (k)
Z32 Z21 xW = Z21 xZ32 Z21 x ⊗ ∂31 (m).
k<b
the last equality being due to the first observation of Remark VII.6.4 (as well as
to multiplication of divided powers).
Remark VII.6.12 The reader may wonder at the fact that we are restricting
the index of summation, k, to be less than b, and l < c. We do this because in the
applications, we want our divided powers of the Z21 terms to be positive (so that
we remain in a normalized complex).
Resolutions of Weyl modules 219
(a) (b)
Definition VII.6.13 If we denote by V the element Z32 Z21 xW, where W is
(c ) (c )
in the submodule spanned by all terms of the form Z211 x · · · xZ21q x ⊗ m, denote
(b−k) (a−k) (k)
by V the term k<b Z21 xZ32 ∂31 W . The module generated by all terms of
the form V − V will be denoted by Cap(1, 2).
This last is an inductive definition that requires a little bit of explanation. If
(b−k) (a−k) (k)
q = 0, then V is simply k<b Z21 x ⊗ ∂32 ∂31 (m) (where we are assuming
that W is just the element m contained in a tensor product of divided powers).
(c)
If W = Z21 x ⊗ m, then
(b−k) (a−k) (c) (k)
V = Z21 xZ32 Z21 ⊗ ∂31 (m)
k<b
(b−k) (c−l) (a−k−l) (l) (k)
= Z21 xZ21 x ⊗ ∂32 ∂31 ∂31 (m).
k<b l<c
In order to show that we still get a complex if we factor out by the submod-
ule, Cap(1, 2), generated by the relations of the type V − V we have just been
discussing, we must show that these relations are carried into Cap(1, 2) by the
boundary map. Since the boundary map, ∂, is just ∂x + ∂y , and since the only
separator in the terms of the relations is x, we need only show that the relations
are stable under ∂x . To this end we have the following proposition, whose proof
is in Reference [30], proposition 5.1.
Proposition VII.6.14 Let W be an element of the bar module of the form
(c ) (c ) (a) (b)
Z211 x · · · xZ21q x ⊗ m, and let V = Z32 Z21 xW. Denote by V the element
(b−k) (a−k) (k)
k<b Z21 xZ32 ∂31 W . Then we have
(a−b)
∂ V = ∂(V ) − Z32 W .
where
Further, we define
and we set
(l ) (l ) (l ) (k ) (k )
Z3,21 yZ3,22 y · · · yZ3,2p Z2,11 xZ2,12 x
4
w 44 1(π) 2(σ1 ) 3(ρ1 )
· · · xZ2,1 x ⊗ w 44 2(σ2 ) 3(ρ2 )
(kq )
4 (ρ3 )
w 3
Resolutions of Weyl modules 221
is simply
lp
(l ) (l ) (k −α) (l −α) (k )
Z3,21 yZ3,22 y · · · yZ2,11 xZ3,2p Z2,12 x
α=0
4 (π−α)
w 4 1 2(σ1 ) 3(α) 3(ρ1 )
4
· · · xZ2,1q x ⊗ w
(k ) 4 2(σ2 ) 3(ρ2 ) .
4
w 4 3(ρ3 )
p−1
If we set |l | = i=1 li , we see that the exponent on the first Z21 is greater than
t1 + |l |, so that condition is preserved. The remaining ones are easy to check;
we point out that it is here that we use the fact that the index of summation in
identity (VII.6.4) is constrained.
Since the description of the terms of our complex, Definition VII.6.15, seems
a bit arbitrary and cumbersome, let us take a little time to give an alternate
description. To do this we first introduce some more notation. In general we
let Zn,m stand for polarizations from place m to place n, and we are going
to suppress the various separators that come into play (as a rule, there will
be a new separator introduced for each operator Zn,m ; in the case at hand,
it will be Z32 with separator y). For a fixed choice of n, m, ξ and l, we will
denote by
ξ
Zn,m Z n,m
(l)
the homogeneous strand of the bar complex of total degree ξ + l with initial
ξ (0)
term of degree ≥ ξ. For example, Z32 Z 32 is just the complex concentrated
222 Some applications of Weyl and Schur modules
ξ(ξ) (2)
in dimension zero consisting of the element Z32 , while Z32 Z 32 is the
complex
(ξ+1)
Z32 yZ32
(ξ) (ξ+2)
0→ Z32 yZ32 yZ32 → ⊕ → Z32 → 0.
(ξ) (2)
Z32 yZ32
If we simply write
Z (l)
n,m ; l > 0,
this will mean the homogeneous strand of the normalized bar complex of degree
l. Now a typical term of Definition VII.6.15 arises in the following way. We choose
(l)
ξ = t2 +1, and any non-negative integer l. We then form the complex Z32 t2 +1
Z 32
and tensor it with
If one looks at the terms of the outcome, one sees that these are the terms
described in Definition VII.6.15. So our complex, Res ((p1 , p2 , p3 ; t1 , t2 )), may
be described as
The foregoing example illustrates how one goes about constructing resolutions
of Weyl modules in an “easy” case. There are many more examples and tech-
niques given in the often-cited paper, [30], which are beyond the scope of this
monograph. However, in that paper, as well as at the end of this paragraph,
the reader will find an explicit description of all the terms of the resolutions
of Weyl modules associated to any almost skew-shape. A discussion of the
obstacles to describing the boundary maps explicitly is also in that article, as
are alternative approaches that may prove to be more effective in solving this
problem. Here, then, is the explicit, inductive description of the terms of the
resolutions.
Res(p1 , . . . , pn ; t1 , . . . , tn−1 )
= Res(p1 , . . . , pn−1 ; t1 , . . . , tn−2 ) ⊗ D(pn )
t +1 (l )
⊕ n−1
Zn,n−1 Z n,n−1
n−1
ln−1 ≥0
⊗ Res(p1 , . . . pn−2 +tn−2 , pn−1 +tn−1 ; t1 , . . . tn−3 +tn−2 , tn−2 +tn−1 − tn−2 )
..
.
t +1 (l ) t +1 (l ) (l )
n−1
(Zn,n−1 Z n,n−1
n−1 n−2
)⊗(Zn,n−2 Z n,n−2
n−2
) ⊗ · · · ⊗(Zn,1
t1 +1
Z n,1
1
)
ln−j ≥0
⊗ D(pn −1≤j≤n−1 tj ) ,
where tj = tj + lj + 1 for j = 1, . . . , n − 1.
(2) For an almost skew-shape (p1 , . . . , pn ; t1 , . . . , tn−1 ) of positive type k, let
us denote by s the positive integer −(tn−1 + · · · + tn−k ), and by τ, the integer
(tn−1 + · · · + tn−k−1 ). With this notation, the terms of the resolution are
Res(p1 , . . . , pn ; t1 , . . . , tn−1 )
= Res(p1 , . . . , pn−1 ; t1 , . . . , tn−2 ) ⊗ D(pn )
(ln−1 )
⊕ Z n,n−1 ⊗ Res(p1 , . . . pn−2 , pn−1 + ln−1 ; t1 , . . . tn−3 , tn−2 + ln−1 )
ln−1 >0
⊗ D(pn −ln−1 )
(ln−2 ) (ln−1 )
⊕ (Z n,n−2 ) ⊗ (Z n,n−1 )
ln−2 >0,ln−1 ≥0
224 Some applications of Weyl and Schur modules
⊗ Res(p1 , . . . pn−2 +ln−2 , pn−1 +ln−1 ; t1 , . . . , tn−3 +ln−2 , tn−2 +ln−1 −ln−2 )
..
.
(l ) (l ) (l )
τ +1
(Zn,n−k−1 Z n,n−k−1
n−k−1
) ⊗ Z n,n−k
n−k
⊗ · · · ⊗ Z n,n−1
n−1
)
ln−j ≥0
t +1 (l ) (l ) (l ) (l )
⊕ n−k−2
(Zn,n−k−2 Z n,n−k−2
n−k−2 τ+1
)⊗(Zn,n−k−1 Z n,n−k−1
n−k−1 n−k
)⊗Z n,n−k ⊗· · ·⊗Z n,n−1
n−1
ln−j ≥0
..
.
(l ) t +1 (l ) (l )
t1+1
(Zn,1 Z n,1
1
)⊗· · ·⊗(Zn,n−k−2
n−k−2
Z n,n−k−2
n−k−2 τ+1
)⊗(Zn,n−k−1 Z n,n−k−1
n−k−1
)
ln−j ≥0
(l ) (l )
⊗ Z n,n−k
n−k
⊗· · ·⊗Z n,n−1
n−1
..
. l1
..
.
kt−1
.. .
. lt−1
kt
..
. lt
The one illustrated is a t-hook, and the indices ki , li are the lengths of the
horizontals (arms) and verticals (legs), respectively. The arms always include
the left-most box, but except for the top-most arm, exclude the right-most box.
The legs always include the bottom box, but exclude the top-most (as it is already
being counted as part of the arm). We insist that the ki always be positive so
that if, for instance, k1 = 1, this means that we are essentially starting with a
vertical piece, etc. We will denote the t-hook by (k1 , . . . , kt ; l1 , . . . , lt ). Note that
if lt = 0, the skew-hook ends with an arm; we have li > 0, however, if i < t.
Thus, the skew-hook represented by (1, 3; 2, 0) looks like this:
In order to describe the resolution of K(k1 ,...,kt ;l1 ,...,lt ) , we first introduce some
general notation. Then we will apply this to our skew-hooks and indicate why
this yields the sought-for resolution.
Let n be a fixed positive integer, and S = {s1 , . . . , sp } be a strictly increasing
set of integers, with sp = n. Denote by Xn the complex consisting of terms, in
226 Some applications of Weyl and Schur modules
Next we look at the following situation. Suppose we have the complex Xk , and
we have two sets of integers T ⊂ S, with both of them containing k. Then since
( S) ⊂ ( T ) , we clearly have Xk / ( S) mapping surjectively onto Xk / ( T ) .
Now we do not know what the kernel of this morphism is in general, but suppose
that the integer n is in S (n < k), and we let T = S − {n}. Then it is easy to see
that the kernel, K, of the morphism is spanned by the terms Di1 ⊗R · · · ⊗R Diα
of Xk / ( S) having the property that for some β, i1 + · · · + iβ = n. If we set
S = {s1 , . . . , sp , n + t1 , . . . , n + tq } , where sp = n, and tq = k − n = m, then K
is isomorphic to Xn / ( {s1 , . . . , sp }) ⊗R Xm / ( {t1 , . . . , tq }) . Moreover, there
is a map of Xk / ( T ) into Xn / ( {s1 , . . . , sp }) ⊗R Xm / ( {t1 , . . . , tq }) which
is fairly easy to define. Namely, suppose that Di1 ⊗R · · · ⊗R Diα is a term in
Xk / ( T ) . This means that for all β, we have i1 + · · · + iβ = n. Then there is
a β0 with i1 + · · · + iβ0 −1 = n − c < n, and i1 + · · · + iβ0 = n + d > n. We map
the term Di1 ⊗R · · · ⊗R Diα to the term
Di1 ⊗R · · · ⊗R Diβ0 −1 ⊗R Dc ⊗R Dd ⊗R Diβ0 +1 ⊗R · · · ⊗R Diα ,
by diagonalizing Diβ0 , which makes sense since iβ0 = c + d . In fact, the mapping
cone of this map (notice the shift of dimension by 1 here), is precisely Xk / ( S) .
Application to skew-hooks
If we now start with a skew-hook (k1 , . . . , kt ; l1 , . . . , lt ), we get by
Theorem VII.1.4 the exact sequence
We want to show, by induction on t and the number of rows, that the resolution
of (k1 , . . . , kt ; l1 , . . . , lt ) is our complex X|k|+|l| / ( S) where S is as above. By
induction, we know that the resolution of (k1 , . . . , kt−1 , kt + 1; l1 , . . . , lt−1 − 1, lt )
is X|k|+|l| / ( T ) , where T = S − {|k| + |l| − lt − 1} . On the other hand, we see
that the resolution of (k1 , . . . , kt−1 ; l1 , . . . , lt−1 − 1) ⊗R (kt + 1; lt ) is the tensor
product of two complexes which is precisely the kernel of the map of X|k|+|l|/( S)
onto X|k|+|l| / ( T ) . Our discussion in the preceding part of this subsection
completes the proof.
The terms of the Lascoux resolution of any skew-shape are read off from the
determinantal expansion of the Jacobi–Trudi matrix of the shape, with the pos-
ition of the terms of the complex determined by the length of the permutations
to which they correspond. In K-theory notation, the Jacobi–Trudi matrix of
the skew-shape λ/µ = (p1 . . . . , pn ; t1 , . . . , tn−1 ) is the n × n matrix
Dp1 Dp1 +t1 +1 Dp1 + t2 ··· Dp1 +t1
Dp2 − t1 Dp2 Dp2 +t2 +1 ··· Dp2 +t2
Dp3 − t2 Dp3 −t2 −1 Dp3 ··· Dp3 +t3
(J − T ) : ,
.. .. .. .. ..
. . . . .
Dpn − tn−1 Dpn −t2 Dpn −t3 ··· Dpn
D2 D3 D4
D1 D2 D3 ,
D0 D1 D2
and the Lascoux resolution of the Weyl module associated to the partition (2, 2, 2)
looks like this:
D3 ⊗R D3 ⊗R D0 D3 ⊗R D1 ⊗R D2
0 → D4 ⊗R D2 ⊗R D0 → ⊕ → ⊕ → D2 ⊗RD2 ⊗RD2 → 0.
D4 ⊗R D1 ⊗R D1 D2 ⊗R D3 ⊗R D1
The correspondence between the terms of the resolution above, and permuta-
tions, is as follows:
D2 ⊗R D2 ⊗R D2 ←→ identity
D3 ⊗R D1 ⊗R D2 ←→ (12)
D2 ⊗R D3 ⊗R D1 ←→ (23)
D3 ⊗R D3 ⊗R D0 ←→ (123)
D4 ⊗R D1 ⊗R D1 ←→ (213)
D4 ⊗R D2 ⊗R D0 ←→ (13).
Resolutions of Weyl modules 229
X0 = D2 ⊗R D2 ⊗R D2 ;
(1) (2)
X1 = Z2,1 x ⊗R D3 ⊗R D1 ⊗R D2 ⊕ Z2,1 x ⊗R D4 ⊗R D0 ⊗R D2 ⊕
(1) (2)
Z3,2 y ⊗R D2 ⊗R D3 ⊗R D1 ⊕ Z3,2 y ⊗R D2 ⊗R D4 ⊗R D0 ;
(1) (1)
X2 = Z2,1 xZ2,1 x ⊗R D4 ⊗R D0 ⊗R D2 ⊕
(1) (2) (1) (3)
Z3,2 yZ2,1 x ⊗R D4 ⊗R D1 ⊗R D1 ⊕ Z3,2 yZ2,1 x ⊗R D5 ⊗R D0 ⊗R D1 ⊕
(2) (3) (2) (4)
Z3,2 yZ2,1 x ⊗R D5 ⊗R D1 ⊗R D0 ⊕ Z3,2 yZ2,1 x ⊗R D6 ⊗R D0 ⊗R D0 ⊕
(1) (1)
Z3,2 yZ3,2 y ⊗R D2 ⊗R D4 ⊗R D0 ⊕
(1) (1)
Z3,2 yZ3,1 z ⊗R D3 ⊗R D3 ⊗R D0 ;
(1) (2) (1)
X3 = Z3,2 yZ2,1 xZ2,1 x ⊗R D5 ⊗R D0 ⊗R D1 ⊕
(2) (3) (1)
Z3,2 yZ2,1 xZ2,1 x ⊗R D6 ⊗R D0 ⊗R D0 ⊕
(1) (1) (3)
Z3,2 yZ3,2 yZ2,1 x ⊗R D5 ⊗R D1 ⊗R D0 ⊕
(1) (1) (4)
Z3,2 yZ3,2 yZ2,1 x ⊗R D6 ⊗R D0 ⊗R D0 ⊕
(1) (1) (1)
Z3,2 yZ3,1 zZ2,1 x ⊗R D4 ⊗R D2 ⊗R D0 ⊕
(1) (1) (2)
Z3,2 yZ3,1 zZ2,1 x ⊗R D5 ⊗R D1 ⊗R D0 ⊕
(1) (1) (3)
Z3,2 yZ3,1 zZ2,1 x ⊗R D6 ⊗R D0 ⊗R D0 ;
(1) (1) (3) (1)
X4 = Z3,2 yZ3,2 yZ2,1 xZ2,1 x ⊗R D6 ⊗R D0 ⊗R D0 ⊕
(1) (1) (1) (1)
Z3,2 yZ3,1 zZ2,1 xZ2,1 x ⊗R D5 ⊗R D1 ⊗R D0 ⊕
(1) (1) (2) (1)
Z3,2 yZ3,1 zZ2,1 xZ2,1 x ⊗R D6 ⊗R D0 ⊗R D0 ⊕
(1) (1) (1) (2)
Z3,2 yZ3,1 zZ2,1 xZ2,1 x ⊗R D6 ⊗R D0 ⊗R D0 ;
(1) (1) (1) (1) (1)
X5 = Z3,2 yZ3,1 zZ2,1 xZ2,1 xZ2,1 x ⊗R D6 ⊗R D0 ⊗R D0 ,
where the subscripts on the X indicate the dimension in which these terms
(t)
appear. The symbols Za,b are the formal “polarization” operators defined in
preceding subsections, following Reference [30], and the letters x, y, z are the
separator variables also explained in that paper and above.
The boundary map for this complex is obtained by polarizing all the separator
(t)
variables to one. When the separator x disappears between a Za,b and elements
(t)
in the tensor product of divided powers, this means ∂a,b , or the place polarization
operator, applied to that tensor product. The only essentially new terms that
(1) (1)
we have here are the terms that involve Z3,2 yZ3,1 z, or more generally terms of
the form:
(1) (1) (k ) (k )
Z3,2 y Z3,1 z Z2,11 x · · · x Z2,1n−2 x ⊗R D2+1+|k| ⊗R D2+1−|k| ⊗R D0 ,
(of which, in this example, there are not that many). For these we have to use
identities of Capelli type, or some easy variants of them (Reference [30] and
230 Some applications of Weyl and Schur modules
with n ≥ 2.
If n = 2, we have
(1)
Z3,1 v = ∂3,1 (v),
while
(1) (1) (1) (2) (2)
Z3,2 Z3,1 z ⊗ v = −Z2,1 x ⊗ ∂3,2 (v)+Z3,2 y ⊗ ∂2,1 (v).
For n > 2, we have
(1) (k ) (k ) (k +1) (1) (k ) (k )
Z3,1 Z2,11 x · · · xZ2,1n−2 x ⊗ v = −Z2,11 xZ3,2 Z2,12 x · · · xZ2,1n−2 x ⊗ v
k1 + k2 (1) (k +k +1) (k )
+ Z3,2 yZ2,11 2 x · · · xZ2,1n−2 x ⊗ v,
k2 − 1
and
(1) (1) (k ) (k ) (1) (2) (k ) (k )
Z3,2 Z3,1 zZ2,11 x · · · xZ2,1n−2 x ⊗ v = −Z2,1 xZ3,2 Z2,11 x · · · xZ2,1n−2 x ⊗ v
(2) (k +1) (k )
+ (k1 − 1)Z3,2 yZ2,11 x · · · xZ2,1n−2 x ⊗ v.
With these identities in play, we can easily write down the boundary map for
this resolution.
VII.6.5.2 Reduction of non-Lascoux to Lascoux
It is of some interest to see how we can discard the “excess” terms of the
characteristic-free resolution to recover that of A. Lascoux. For example, we want
(2) (2)
to throw away the terms Z2,1 x⊗R D4 ⊗R D0 ⊗R D2 and Z3,2 y⊗R D2 ⊗R D4 ⊗R D0 .
(2)
If we look at the image of a term Z2,1 x ⊗ v, where v ∈ D4 ⊗ D0 ⊗ D2 , we
(2) (2)
see that it is ∂21 (v) = 1 2
2 ∂21 (v). So, the image of Z2,1 x ⊗ v is the same as
1 (1) (1)
the image of ⊗ 2 Z2,1 x ∂21 (v). Hence
in characteristic zero we can rig up the
boundary map taking this into account. Obviously the same kind of thing holds
Resolutions of Weyl modules 231
(2)
for the term Z3,2 y ⊗R D2 ⊗R D4 ⊗R D0 that is, its image is the same as that of
1 (1) (1)
2 Z3,2 x⊗ ∂32 (v). Thus, any term that is sent by the “full” boundary map into
one of the “redundant” terms above, should now be sent into the non-redundant
term instead. For example, the following terms should now be sent as indicated:
(1) (2) (1) (2) (1) (1)
Z3,2 yZ2,1 x ⊗ v → Z3,2 y ⊗ ∂2,1 (v) − Z2,1 x ⊗ ∂3,1 (v) − 12 Z2,1 x ⊗ ∂2,1 ∂3,2 (v)
and
(1) (1) (1) (1) (2) (1)
Z3,2 yZ3,1 z ⊗ v → Z3,2 y ⊗ ∂3,1 (v) − Z2,1 x ⊗ ∂3,2 (v) − 12 Z3,2 y ⊗ ∂3,2 ∂2,1 (v).
With this modification, the terms of the type
(1) (1)
Z2,1 xZ2,1 x ⊗R D4 ⊗R D0 ⊗R D2
and
(1) (1)
Z3,2 yZ3,2 y ⊗R D2 ⊗R D4 ⊗R D0
are automatically sent to zero. As a result, when we go to the last term that
counts in this complex, we see that under the boundary map in the big complex
we get
(1) (1) (1) (1) (1) (1) (2)
Z3,2 yZ3,1 z Z2,1 x ⊗ v → Z3,2 yZ3,1 z ⊗ ∂2,1 (v) + Z3,2 y Z2,1 x ⊗ ∂3,2 (v)
(1) (1) (2) (1) (1) (2)
− Z3,2 yZ3,2 y ⊗ ∂2,1 (v) − Z2,1 xZ2,1 x ⊗ ∂3,2 (v),
so that under the modified boundary, we can simply define the boundary map
on this term to be
(1) (1) (1) (1) (1) (1) (2)
Z3,2 yZ3,1 z Z2,1 x ⊗ v → Z3,2 yZ3,1 z ⊗ ∂2,1 (v) + Z3,2 y Z2,1 x ⊗ ∂3,2 (v).
Of course it remains to prove that in characteristic zero this is exact; we will
now indicate how to do this.
VII.6.5.3 Question of exactness
Here, we indicate more schematically how we modify the maps to take advantage
of divisibility in Q. Divide the terms of our big complex into the sum of those
of the Lascoux complex, and the others. That is, we look at the terms of the
complex as:
X0 = A0
X1 = A1 ⊕ B1
X2 = A2 ⊕ B2
X3 = A3 ⊕ B3
Xj = Bj for j = 4, 5,
where the Ai are the sums of the “Lascoux” terms, and the Bl are the sums of
the others.
232 Some applications of Weyl and Schur modules
such that
(C2 ) δB2 A1 + σ1 δB2 B1 = {δA2 A1 + σ1 δA2 B1 }σ2 .
We define this map as follows:
(1) (1)
Z2,1 x Z2,1 x ⊗ v −→ 0
(1) (1)
Z3,2 y Z3,2 y ⊗ v −→ 0
(1) (3) (1) (2)
Z3,2 y Z2,1 x ⊗ v −→ 13 {Z3,2 y Z2,1 x ⊗ ∂2,1 (v)}
(2) (3) (1) (2) (1) (1) (2)
Z3,2 y Z2,1 x ⊗ v −→ 13 {Z3,2 yZ2,1 x ⊗ ∂3,1 (v)−Z3,2 yZ3,1 z ⊗ ∂2,1 (v)}
(2) (4) (1) (1) (3)
Z3,2 y Z2,1 x ⊗ v −→ − 12 {Z3,2 y Z3,1 z ⊗ ∂2,1 (v)}
where the element v is in the appropriate tensor product of divided powers as
given in the description of the large complex.
Remark VII.6.18 The map σ2 defined above satisfies the condition (C2 ).
Proof Trivial. 2
Remark VII.6.19 We have exactness at A1 .
Proof This is just a diagram chase. 2
Using σ2 we can now also define
∂3 : A3 → A2
by
∂3 = δA3 A2 + σ2 δA3 B2 .
Not too surprisingly, we have
Remark VII.6.20 ∂2 ∂3 = 0.
Of course, what we need now is a map σ3 : B3 → A3 similar to the maps σ
above, that is, satisfying
(C3 ) δB3 A2 + σ2 δB3 B2 = {δA3 A2 + σ2 δA3 B2 }σ3 .
We define such a σ3 as follows:
where the element v again is taken in the appropriate tensor product of divided
powers as described earlier.
Remark VII.6.21 The map σ3 defined above satisfies the condition (C3 ).
Proof Trivial. 2
In this appendix, we furnish the proofs of the theorems stated in Sections VI.3
and VI.8 on letter-place methods. More precisely, Sections A.1 and A.2 give a
complete proof of Theorem VI.3.2 which involves straightening considerations
(Section A.1) as well as a combinatorial procedure known as the Robinson–
Schensted–Knuth correspondence (Section A.2). In Section A.3 we deal in a less
detailed way with the proofs of Theorems VI.3.3 and VI.3.4, mainly showing what
modifications are necessary to adapt the methods of proof of Theorem VI.3.1 to
the other situations. In Section A.4, we discuss the modifications necessary for
the proof of Theorem VI.8.4.
Since the set of tableaux is finite, this will show that we can express any double
tableau as a linear combination of such that are standard in the places. We then
do the same for the letters. As this process cannot go on forever, each double
tableau must eventually be a linear combination of standard double tableaux;
that is, the standard ones generate.
At the cost of being ultra pedantic, we shall denote a double tableau, T , by a
triple: T = (λ; L, P ), to denote the diagram, the “letter,” and the “place” parts
of the tableau. We stress that the diagram is always to be that of a partition
in this case, that is, the lengths of the rows are decreasing from top to bottom.
In Subsection VI.1.1, we described a well-known partial order on partitions: we
said that λ > λ if the first row of λ (from the top) which differs in length
from the corresponding row of λ, is longer than that of λ. We also defined in
Subsection VI.1.2, a quasi order on single tableaux. Using these orderings, we
now define a quasi ordering of our double tableaux.
Definition A.1.1 Given tableaux, T = (λ; L, P ) and T = (λ ; L , P ), we
define (λ ; L , P ) ≤ (λ; L, P ) if λ ≥ λ, Lij ≥ Lij for all i, j, and Pij ≥ Pij
for all i, j. We then say that (λ ; L , P ) < (λ; L, P ) if (λ ; L , P ) ≤ (λ; L, P ),
and either λ > λ or λ = λ and either Lij > Lij for some i, j, or Pij > Pij for
some i, j.
To give an example, and also to see how the general “straightening” proceeds,
consider an elementary, yet prototypical situation:
w1 1(a1 ) 2(b1 )
T = ,
w2 1(2) 2(b2 )
where w1 and w2 are words of degrees a1 + b1 and 2 + b2 respectively. (We are
now dropping the cumbersome designation of a tableau as a triple.)
Our problem in the “place” section of the tableau is that we have 1 in the
bottom row, so that no matter what the value of a1 , that section is not standard.
Recall that
T = w1 (a1 )w2 (2) ⊗ w1 (b1 )w2 (b2 )
where the letters in parentheses indicate the degrees of the wi under appropriate
diagonalization. (We will continue to use this notation to indicate that we have
diagonalized our terms in the indicated degrees.)
Consider now:
(a1 +2) (b1 )
w 1 w2 (2) 1 2
T = ,
w2 (b2 ) 2(b2 )
w1 1(a1 +1) 2(b1 −1)
T = ,
w2 1 2(b2 +1)
w1 1(a1 +2) 2(b1 −2)
T = .
w2 2(b2 +2)
Theorem VI.3.2, Part 1 239
• It is extremely important to observe at this point that the diagram for T has
changed shape, but it is bigger than that of our original T . The shapes of T and
T are the same as our original, the letter part of the tableau has not changed
(that is, L = L = L), but we have both P and P which are different from
P . In both cases, though, we have decreased these tableaux (in the quasi order).
Notice that if we were to “straighten” our letter tableau similarly, the changes
would either lead to a bigger diagram, in which case we go down in the order,
or the places would remain unchanged, and the letter tableaux that intervened
would be lower in the quasi order.
• When checking this type of calculation, it is most often convenient to
assume that the words, wi , are simply divided powers. That is, we might assume
that w1 = x(a1 +b1 ) and that w2 = y (2+b2 ) . This makes keeping track of the
diagonalizations much easier.
The point of this exercise is to show that when we get rid of an instance of
non-standardness, the tableaux that emerge in the process are all less, in the
quasi-order we introduced, than the tableau we started with. In the case of T ,
this is due to the fact that the diagram of T is properly larger than that of T ;
in the other two cases, the “letter” side of the tableau has not changed, but the
“place” side has properly decreased. Of course, the tableau, T , is manifestly
non-standard (the others may or may not be, depending on the relative values
of a1 , b1 and b2 ).
To see how a greater number of places affect the straightening procedure, we
will look at one more example. To this end, consider:
w1 1(a1 ) 2(b1 ) 3(c1 )
T = .
w2 1 2(b2 ) 3(c2 )
Then if we set:
w1 w2 (1) 1(a1 +1) 2(b1 ) 3(c1 )
T = .
w2 (b2 + c2 ) 2(b2 ) 3(c2 )
w1 1(a1 +1) 2(b1 −1) 3(c1 )
T = ,
w2 2(b2 +1) 3(c2 )
and
w1 1(a1 +1) 2(b1 ) 3(c1 −1)
T = ,
w2 2(b2 ) 3(c2 +1)
we have
T = T − (b2 + 1)T − (c2 + 1)T .
240 Appendix for letter-place methods
Again we see that we have eliminated the offending 1 in the bottom row, and
we have “replaced” T by the tableaux T , T and T which are strictly lower in
our quasi-order than T .
As these two examples show, it is enough to work either on the letter or place
side of the tableau to push us further toward standardness. These examples are
prototypical in the sense that it is enough to work on straightening two adjacent
rows, for if we have a tableau that is not standard, this is because it is not
strictly increasing in the columns (since we start with row-standardness). But
then a violation of strict increase must also occur in two adjacent rows. This is
the reason that we can focus on two-rowed double tableaux such as:
w1 z11 z12 z13 · · · z1k
T =
w2 z21 z22 z23 · · · z2l
where we write zij to represent the integers from 1 to n, with possible repeats.
Suppose that the first violation of strict increase occurs in the ith column, that
is, z11 < z21 , . . . , z1i−1 < z2i−1 , but z1i ≥ z2i . (The ith column could be the first,
as it was in the above examples.) Next, suppose that z2i = · · · = z2t < z2t+1
(possibly t = i), and consider the sum of tableaux:
w1 w2 (t) z11 · · · z1i−1 z21 · · · z2t z1i · · · z1k
T = .
w2 (l − t) z2t+1 · · · z2l
Now, for each subset J of U = {z21 , . . . , z2t , z1i , . . . , z1k }, of order 0 < s < t,
let J be the complement of J in U , let ZJ be the row tableau of that set, and ZJ
the corresponding row tableau of J. We let I stand for an arbitrary subset of U of
order precisely t, other than the subset {z21 , . . . , z2t } itself, I its complementary
subset, and ZI , ZI as for J. Set
w1 w2 (t − s) z11 · · · z1i−1 ZJ
TJ = ,
w2 (l − t + s) ZJ z2t+1 · · · z2l
and cJ equal plus or minus the product of binomial coefficients that would be
appropriate due to the multiplication in the divided power algebra, of ZJ with
z2t+1 · · · z2l (see the foregoing examples to get a more concrete picture of this
description). Finally, let
w1 z11 · · · z1i−1 ZI
TI = ,
w2 ZI z2t+1 · · · z2l
and cI equal plus or minus the corresponding product of binomial coefficients.
Then it is tedious, but straightforward to prove that
T = ±T + cJ TJ + cI TI .
J I
from that of T . However, since our sets I must have t elements, and none can be
the set {z21 , . . . , z2t } itself, at least one of these elements must remain in the top
row, and one of the elements z1u > z2i must come down into the bottom row.
Thus, the resulting tableaux in this case are all less in the quasi order than the
original.
One may ask, when dealing with the tableaux other than the TI what we do
about keeping these tableaux within the class we are considering, namely, the
lengths of the rows decreasing. In order to keep to this recipe, we simply push
the top row of the two up as far as it has to go, and the bottom one down to
where it has to go to make it into a legitimate shape. But this still makes the
shape lower in our quasi order than the original.
To assign the next tableau, P (β), we will make a slight modification of the
usual R–S–K, and make use of the fact that we are looking very distinctly at
D3 ⊗ D4 , namely, a two-fold tensor product of divided powers of designated
degrees. This means that we have only two places to deal with, namely 1 and
2, and we want P (β) to be of the same shape as L(β), standard, and filled with
the places 1 and 2. Now the first part of the construction of our tableau involved
using the first three terms of the sequence, β, so that the first three boxes of the
first row should be filled in with the place, 1. The second row was next produced
by bumping, so we put the place 2 in that box of the second row. The next
three entries in the first row were produced by inserting entries from the second
factor, so we fill them in with 2, and the resulting tableau we get is P (β) =
1 1 1 2 2 2
.
2
On the other hand, given the pair of tableaux, L(β) and P (β), we can recon-
struct the sequence β. We look at the tableau, P (β), and peel off the highest
place from the highest row first, with its corresponding letters in L(β). This
tells us that in our second factor, we had x2 x2 x3 , and we still are left with the
x x1 x2 1 1 1
pair of tableaux: 1 and . This means that we still have a
x3 2
term from the second factor, and it was obtained by bumping from a single-rowed
tableau with three entries. The only way that bumping could have happened was
for x3 to have been in the upper row previously, and to have been bumped by x2
(for if an element of the first row had been bumped by x1 , that element would
have been x2 , and we would have had x2 in the second row). Therefore, the
full second factor must be x2 x2 x2 x3 , and our first factor is what is left, namely
x1 x1 x3 .
To explain the bumping procedure that leads to the construction of the
first tableau, let us consider the general situation of an ordered set, S =
{s1 , s2 , . . . , st }, and a sequence β = si1 si2 . . . sik formed from these elements.
To the element si1 we attach the tableau consisting of one box, and the one
entry, si1 . If si2 ≥ si1 , then we attach the one-rowed tableau consisting of si1
and si2 ; if si2 < si1 , we attach the two-rowed tableau having si2 in the top row,
and si1 in the second row. Suppose we have attached by this procedure a stand-
ard tableau, λ, to the first l elements of the sequence. We then take the element
sil+1 and try to add it to λ. If it exceeds (in the weak sense) every element in
the top row of λ, then we stick it onto the end of that row. If it does not, then
we look for the first element of the first row that is strictly greater than it, and
replace it by sil+1 . This leaves us with the exiled element of the first row, and
we take it and look at the second row. If it fits, we add it on to the second row,
otherwise bump as before and continue in this way. By this procedure, we arrive
at the tableau, L(β), associated to the sequence, β.
What if, one may ask, the first two rows of λ were of the same length and
the bumped element exceeded all the elements of the second row? Would not
sticking it onto the second row produce an inadmissible shape? But we have
Theorem VI.3.2 Part 2 243
assumed that the tableau, λ, is standard, and so we see that this situation cannot
arise.
To associate a “place” tableau to the sequence β, we have to be given a bit
more information along with the sequence (as when we saw in our example that
we were dealing with two factors—hence two places—and subsequences of fixed
length). So let us assume that our integer k = k1 + · · · + kn , and that our
sequence β has the property that the first k1 elements are increasing, the next
k2 are increasing, and so on. We want to assign to our β a standard place tableau
having n places. Notice that since the first k1 elements are increasing, they all
fit into a single-row tableau with k1 boxes. We record this by attaching a single-
rowed tableau with k1 boxes all filled in with the place 1. We then run through
the next k2 elements, keeping track of all the new boxes created in the L(β)
construction by labeling them with the place 2. Notice that if elements from the
second strand bump elements from the first row, the elements they bump grow
in size, so that these elements all go into the second row (they bump no more).
Also, once an element from this second strand gets placed in the first row, all
the others do. This means that the place tableau so far associated has 1’s in the
first row, and 2’s in the first and second rows. Obviously the number of 2’s in
the second row cannot exceed the number of 1’s in the first row (namely, k1 ).
We then proceed with the third strand placing 3’s, and so on.
Rather than spend more words on this description, a not too trivial example
may be in order. For simplicity, we will use numbers for letters as well as places;
we believe that this should cause no confusion.
the 5 had been the bumper, it would have already fit nicely onto the first row.
So, it must have been the 3 that bumped, which means that the third factor was
3446, and that 3 bumped into the two-rowed tableau having 122335 in the top
row and 23446 in the second. The corresponding place tableau now has only the
places 1 and 2 with all the entries labeled 1 in the top row and all those labeled
2 in the second. Now the previous tableau had six entries in the top row, and
only four in the second. The only way we could have arrived at the current stage
is if we had bumped a 5 into the tableau having 122346 in the top row, and 2344
in the second. And so forth. In this way we reconstruct our original sequence of
departure.
As for the case of negative letters and positive places, the proof that the
appropriate double standard tableaux generate is just a mild modification of the
above. The proof of linear independence involves a modified R–S–K, namely, in
this case, you build the letter tableau by rows, and again label the new boxes by
the place corresponding to the element that you bump in. To illustrate, we will
use the same element we used above, namely x2 ∧ x3 ∧ x5 ⊗ x1 ∧ x3 ⊗ x2 ∧ x4 ⊗
x3 ∧ x5 ∧ x6 in Λ3 F ⊗ Λ2 F ⊗ Λ2 F ⊗ Λ3 F. Our R–S–K applied to this yields the
double tableau:
x1 x2 x3 x5 x6 1 1 1 4 4
x2 x3 x4 2 2 3 .
x3 x5 3 4
The reader should be careful not to confuse this double standard tableau
with the double tableau that actually equals the element x2 ∧ x3 ∧ x5 ⊗ x1 ∧
x3 ⊗ x2 ∧ x4 ⊗ x3 ∧ x5 ∧ x6 in Λ3 F ⊗ Λ2 F ⊗ Λ2 F ⊗ Λ3 F. For this element
246 Appendix for letter-place methods
is simply
x2 x3 x5 1 1 1
x1 x3 2 2
,
x2 x4 3 3
x3 x5 x6 4 4 4
For the case of the symmetric algebra, the proof that the appropriate double
standard tableaux generate presents no new feature. As for their linear inde-
pendence, the R–S–K builds the letter part of the tableau by columns, and the
place part of the tableau in the usual way. To take an example, consider the
element x22 x3 x4 ⊗ x31 x2 ⊗ x2 x24 ∈ S4 F ⊗ S4 F ⊗ S3 F . The corresponding standard
double tableau is
x1 x2 1 2
x1 x2 1 2
x1 x3 1 2
x2 x4 1 2 .
x2 3
x4 3
x4 3
where si + ti = ki and the xij , yij are elements in G, F, respectively such that for
all i, j, xij ≤ xij+1 , yij < yij+1 . We now must define the R–S–K that will assign
to such an element a double standard tableau. But this modification is simple:
we use the procedure that we used before, but this time we allow repeats in the
x entries, and no repeats in those of y. An example here will tell the whole story.
Theorem VI.8.4 247
[69] C. Peskine and L. Szpiro, Dimension projective finie et cohomologie locale, Publ.
Math. IHES 42 (1973), 47–119.
[70] P. Pragacz and J. Weyman, On the construction of resolutions of determinantal
varieties: a survey, in Séminaire d’algèbre Paul Dubreil et Marie-Paule Malliavin,
37ème Année (Paris, 1985), Lecture Notes in Mathematics 1220, Springer-Verlag
(1986), pp. 73–92.
[71] —— and ——, On the generic free resolutions, J. Algebra 128 (1990), 1–44.
[72] G.A. Reisner, Cohen-Macaulay quotients of polynomial rings, Adv. Math. 21
(1976), 30–49.
[73] J. Roberts and J. Weyman, A short proof of a theorem of M. Hashimoto, J.
Algebra 134 (1990), 144–156.
[74] P. Roberts, The vanishing of intersection multiplicities of perfect complexes, Bull.
Am. Math. Soc. 13 (1985), 127–130.
[75] ——, Intersection multiplicities in commutative algebra, in Algebra colloquium
(Rennes, 1985), Univ. Rennes I, Rennes 1985, pp. 25–39.
[76] ——, Multiplicities and Chern Classes in Local Algebra, Cambridge Tracts in
Mathematics No. 133, Cambridge University Press, Cambridge (1998).
[77] P. Samuel, On unique factorization domains, Illinois J. Math. 5 (1961), 1–17.
[78] J.-P, Serre, Algèbre locale. Multiplicités. Cours au Collège de France, 1957–1958,
rédigé par Pierre Gabriel. Seconde édition, 1965. Lecture Notes in Mathematics 11,
Springer-Verlag, Berlin-New York (1965).
[79] W. Soergel, Conjectures de Lusztig, Astérisque 237 (1996), 75–85.
[80] R.G. Swan, Algebraic K-Theory, Lecture Notes in Mathematics 76, Springer-
Verlag, New York (1968).
[81] B.D. Taylor, Generalized straightening laws for products of determinants, Ph.D.
Thesis, MIT, (1997).
[82] ——, Compressed straight tableaux and a distributive lattice of representations, J.
Combin. Theory A 91 (2000), 598–621.
[83] ——, A straightening algorithm for row-convex tableaux, J. Algebra 236 (2001),
155–191.
[84] J. Towber, Two new functors from modules to algebras, J. Algebra 47 (1977),
80–109.
[85] J. Weyman, On the structure of free resolutions of length 3, J. Algebra 126 (1989),
1–33.
[86] ——, Cohomology of Vector Bundles and Syzygies, Cambridge Tracts in Mathem-
atics No. 149, Cambridge University Press, Cambridge, 2003.
[87] O. Zariski and P. Samuel, Commutative Algebra (Vols 1 and 2), Springer-Verlag,
Berlin (1979).
INDEX
Koszul complex, 39
Krull Principal Ideal Theorem, 50 Quillen-Suslin theorem, 15, 114