Arun Ram
Department of Mathematics and Statistics
University of Melbourne
Parkville, VIC 3010 Australia
aram@unimelb.edu.au
Last update: 28 June 2013
Schubert Polynomials (1)
Let
so that
For each permutation the Schubert polynomial
is defined to be
where as usual is the longest element of
(4.2)
Let Then
In particular,
Proof.
From (2.7) we have
Now
and
by (1.6). Hence
is equal to
if
and is zero otherwise.
(4.3)
(i)
(ii)
For each
is a non-zero homogeneous polynomial in
of degree of the form
summed over such that
(i.e.,
for each and
(iii)
is symmetrical in
if and only if
(iv)
If is the last descent of (i.e., if
and
then
and
Proof.
(i) That is clear from the definition (4.1).
Also by (2.11) we have
(ii) The operator
lowers degrees by
Hence
is homogeneous of degree If now
is such that
then by (2.1)
is a linear combination of monomials such that
if
and
so that Hence the linear span
of the monomials is mapped into itself by
each
and hence by each
Hence for each
(iii) is symmetrical in and
if and only if
that is to say if and only if
which by (4.2) is equivalent to
(iv) is symmetrical in
by (iii) above, but does not contain hence does not contain any of
Remark. We shall show later (4.17) that the coefficients in (4.3)(ii) are always non-negative integers.
(4.4)
For
we have
Proof.
By (4.3), is a homogeneous symmetric polynomial of degree
in
hence is equal to
for some integer But
by (4.2) and (4.3)(i), hence
(4.5) (Stability)
Let and let
be the embedding. Then
for all
Proof.
We may assume that Let
be the longest element of
then
where is the longest element of and hence
(because
hence
and so on.)
From (4.5) it follows that is a well-defined polynomial for each permutation
If and
we denote by the permutation
in We have then
where is the identity element of
Proof.
We shall make use of the following fact: if is a polynomial in
and
for all
then is a constant. For
for some and is symmetric in
because
To prove (4.6) we proceed by induction on
If
then
and both sides of (4.6) are equal to Let
By the remark above, it is enough to show that
for each
Suppose first that Then
because
by (4.2). Hence we have
if
and if then
which is zero by the inductive hypothesis.
Likewise, if we have
and so again
by the inductive hypothesis.
Finally, if we have
because
and therefore kills and
moreover,
because
Hence
and the proof is complete.
For certain classes of permutations there are explicit formulas for We consider first
the case where is dominant, of shape (so that the diagram of coincides with the diagram of
(4.7)
If is dominant of shape then
Proof.
We use descending induction on where
The result is true for
by (4.3)(i), since is dominant of shape
Suppose
and is dominant of shape
Then and Let
be the largest integer such that
for and let
Then is dominant of length
and
where is the vector whose component is
and all other components zero. Hence we have
because
Conversely, every monomial (where is a partition) occurs as a Schubert polynomial,
namely as where is the permutation with code
Suppose next that is Grassmannian, with descent at
(4.8)
If is Grassmannian of shape then
is the Schur function
where is the unique descent of and
Proof.
We may assume that (by (4.3)(i),
Then and the code of is
so that
Let be the longest element of
Then
is dominant of shape where
and
Hence
by (4.2), (4.7) and (2.11).
Conversely, every Schur function
(where is a partition of length occurs as a Schubert polynomial,
namely as where is the permutation with code
More generally, let be vexillary with shape
(where and flag
(Chapter I). Then is a multi-Schur function (Chapter III), namely
where
for each
Proof.
The idea is to convert systematically into a dominant permutation. Recall ((1.23), (1.24)) that if
and for some
then
and
As in Chapter I let
where
(and each and let
where
Consider first the terms equal to in the sequence
They occupy the positions
We shall use to move them all to the left until they occupy the first
positions, by multiplying on the right by
Let In the code of
the first entries will be equal to
the shape of is
and it follows from the description (1.38) of vexillary codes that the terms equal to in the sequence
will occupy the positions
The next step is to move those to the left until they occupy the positions
by multiplying on the right by
Let
the code of starts off with entries to
then entries equal to
the shape of is
and the terms equal to in the sequence
will occupy the positions
We continue in this way; at the stage we define
where
and has shape
where
by (1.36). Notice also that
by (1.37).
Finally we reach
which is dominant with shape (and code)
We have
and
so that
and therefore, since
by (4.2). Now by (4.6) and (3.5') we have
where
Hence by repeated use of (3.10) we obtain
by virtue of (3.6). If we now operate with we shall obtain in the same way
and so finally
Remarks. 1. As in Chapter I, let
be the conjugate partition, so that
and therefore
by (1.41), where
is the flag of Thus
2.
The result (4.9) admits a converse. If
as above, every non-zero multi-Schur function
that satisfies the conditions of the duality theorem (3.8''), namely
is the Schubert polynomial of a vexillary permutation, namely the permutation with shape and flag
This follows from (1.38) and (4.9), since the conditions (1) on the flag coincide with those of (1.37). (The conditions (1.36), namely
ensure that the multi-Schur function does not vanish indentically.)
Let denote the additive subgroup of
spanned by the monomials
(4.11)
The Schubert polynomials
form a of
Proof.
By (4.3) each lies in If
is a linear dependence relation, then by homogeneity we have
for each and by operating on (1) with we see
that Hence the
are linearly independent and hence form a of
It follows that each monomial
can be expressed in the form
with rational coefficients by operating on (2) with
we have
and hence the are integers.
From (4.11) it follows that
(4.12)
The
form a of
Proof.
Let be a monomial in
Then for sufficiently large
hence is a linear combination of the
For each let
denote the set of all permutations such that
or equivalently such that the code of has length
(4.13)
The
form a of
Proof.
By (4.3)(iii) we have
Let be the
of the
If choose
by (4.12) we can write as a linear combination of Schubert polynomials, say
where there is at least one term with and
Hence for some we have
and since we obtain from (1) a nontrivial linear dependence relation
among the Schubert polynomials, contradicting (4.12). Hence
which proves (4.13).
Let be the homomorphism defined by
In other words,
is the constant term of
for each polynomial The expression of
in terms of Schubert polynomials is then
Proof.
By (4.13) and linearity, it is only necessary to verify this formula when is a Schubert polynomial
and then it follows from (4.2) and (4.3)(ii) that
is equal to when and is zero otherwise.
(4.15)
Let
be a homogeneous linear polynomial, and let be a permutation. Then
where is the transposition that interchanges and
and the sum is over all pairs such that
Proof.
The polynomial is homogeneous of degree
and hence by (4.14) we have
summed over of length Now by (2.13)
summed over such that
It follows that
if and is zero otherwise.
In particular:
summed over transpositions such that
where or
according as or
(4.15'') (Monk's formula)
summed over transpositions such that
and
Remark. As pointed out by A. Lascoux, Monk's formula (4.15'') (which is the counterpart of Pieri's formula in the theory of Schur functions)
characterizes the algebra of Schubert polynomials.
We shall apply (4.15') in the following situation. Suppose that is the last descent of
so that
and
Choose the largest such that
and let
Then from (4.15') applied to we have
summed over all permutations
where and
Hence
and for
Let us arrange the permutations of a given length in reverse lexicographical ordering, so that
if
then precedes if and only if for some
we have
For this ordering there is a first element, namely the permutation
We have proved
(4.16)
For each permutation the Schubert polynomial can be expressed in the form
where is the last descent of
and each in the sum precedes in the reverse lexicographical ordering.
From (4.16) we deduce immediately that
(4.17)
For each permutation is a polynomial in
with positive integral coefficients.
Proof.
For we may assume, as inductive hypothesis, that (4.17) is true for all permutations such that either
or
and precedes in the reverse lexicographical ordering; and then (4.16) shows that the result is true for
(The permutation
has code hence is dominant with Schubert polynomial
by (4.7)·)
Now fix integers such that
and let so that
By (4.12) we can express
uniquely in the form
summed over and
(4.19)
The coefficients in (4.18) are non-negative integers.
Proof.
We proceed by induction on Suppose first that
and that
so that
Then there exists such that
From (4.18) we conclude that
hence is equal to and therefore we have
and
By the inductive hypothesis, we conclude that if
It remains to consider the case Let
be the homomorphism for which
if and
if From (4.18) we have
Let be the last descent of If then
and hence
so that is equal to if
and is zero otherwise. If we deduce from (4.16) that
□
Kohnert's algorithm
Let D be a "diagram", which for present purposes means any finite non empty set of lattice points
(i,j) in the positive quadrant
(i≥1,j≥1).
Choose a point p=(i,j)∈D which is rightmost in its row,
and suppose that not all the points (1,j),…,(i-1,j)
directly above p belong to D. If h is the largest integer less than
i such that (h,j)∈D, let
D1 denote the diagram obtained from D by replacing
p=(i,j) by
(h,j). We can then repeat the process on
D1, by choosing the rightmost element in some row, and obtain a diagram
D2, and so on. Let K(D)
denote the set of all diagrams (including D itself) obtainable from D by a sequence of such moves.
Next, we associate with each diagram D a monomial
xD=∏i≥1xiai
where ai is the number of elements of D in the
ith row, i.e., the number of j such that
(i,j)∈D.
With this notation established, Kohnert's algorithm states that
(4.20)
For each permutation w we have
𝔖w=∑D∈K(D(w))xD
where D(w) is the diagram (1.20) of w.
Example.
If w=(1432),K(D(w)) consists of the diagrams
and
𝔖w=x22x3+x1x2x3+x1x22+x12x3+x12x2.
A proof of a related algorithm by N. Bergeron is given in the Appendix to this chapter. The present status of (4.20) is that it is true for w
vexillary [Koh1990], but open in general.
The shift operator
Let f∈Pn and let m≥n. Then
(4.21)τf=τmf=∂1…∂m(x1…xmf)=π1…πm(f)
is independent of m, because πmf=f
if f is symmetrical in xm and
xm+1, and in particular if f does not contain
xm,xm+1.
The operator τ:Pn→Pn+1
is called the shift operator. For example, we have
(4.23)
Let α∈ℕn and
0≤p1≤…≤pn.
Then
τsα(Xp1,…,Xpn)=sα(Xp1+1,…,Xpn+1).
Proof.
Since τ=π1π2…πpn,
this follows from (3.10).
□
(4.24)
We have
∂iτr=0
for 1≤i≤r.
Proof.
By (4.12) it is enough to show that ∂iτr𝔖w=0
for all permutations w, and this follows from (4.22) and (4.2).
□
For each n≥1 let ρn:P∞↦Pm
be the homomorphism defined by
ρn(xi)={xiifi≤n,0ifi>n.
(4.25)
Let w0(n) be the longest element of
Sn. Then
πw0(n)(f)=ρnτn(f)
for all f∈Pn.
Proof.
By linearity we may assume that f=xα where
α∈ℕn. Since
xα=sα(X1,…,Xn)
by (3.5'), we have
τn(xα)=sα(Xn+1,…,X2n)
by (4.23), and hence
ρnτn(xα)=sα(Xn,…,Xn)
which is equal to πw0(n)(xα)
by (2.16')·
□
Transitions
A transition is an equation of the form
T(w,r)𝔖w=xr𝔖u+∑v∈Φ𝔖v
where r≥1,w and u are permutations and
Φ is a set of permutations. It exists only for certain values of r, depending on
w. An example is (4.16), in which r is the last descent of w.
By (4.15') we have
xr𝔖u=∑tσ(t)𝔖ut
summed over transpositions t=tir such that
ℓ(ut)=ℓ(u)+1,
where σ(t) is the sign of i-r.
So for T(w,r) to hold there must be exactly one
j>r such that
(1)ℓ(utrj)=ℓ(u)+1,(2)w=utrj.
Consider the graphs G(w) and
G(u) of w and u.
They differ only in rows r and j:
G(w)G(u)
By (1.10) the relation (1) above is equivalent to A∩G(u)=∅,
where A is the open region indicated in the diagram. Moreover, j is the only integer >r
such that u(j)>u(r)
and A∩G(u)=∅,
and this will be the case if and only if (A∪B∪C)∩G(u)
is empty. Since (A∪B∪C)∩G(u)=(A∪B∪C)∩G(w),
it follows that
(4.26)
There is a transition T(w,r) if and only if
(A∪B∪C)∩G(w)=∅.
From (4.26) it follows that if T(w,r) exists we must have
w(r)>w(r+1),
i.e., r must be a descent of w. Hence
d0(w)≤r≤d1(w)
where d0(w) (resp.
d1(w)) is the first (resp. last) descent of
w. (In terms of the code c(w),d0(w)
is the first descent of the sequence c(w), and
d1(w) is the largest i such that
ci(w)≠0.)
In general, not all descents of w will give rise to transitions, but the last descent always does, by (4.16).
Consider next the set Φ=Φ(w,r)
of permutations that feature in T(w,r). Each
v∈Φ is of the form v=utir
with i<r and ℓ(v)=ℓ(u)+1(=ℓ(w)).
Again by (1.10), this means that
G(w)G(u)G(v)
A′∩G(w)
is empty, where A′ is the open region indicated in the diagram above.
The element v=utir of
Φ for which i is maximal is called the leader of
Φ. Thus v∈Φ is the leader if and only if
(4.27)(A′∪B′)∩G(w)=∅.
Remark (4.28).
The set Φ will be empty if and only if there is no i<r such that
w(i)<w(j).
We can always avoid this possibility by replacing w by 1×w. If
Φ(w,r) is not empty, then
v↦1×v is a bijection of
Φ(w,r) onto
Φ(1×w,r+1).
The condition (4.26) is stable under reflection in the main diagonal, which interchanges G(w)
and G(w-1). Hence
(4.29)
The transition T(w,r) exists if and only if
T(w-1,s)
exists, where s=w(j). Moreover we have
Φ(w-1,s)=Φ(w,r)-1
so that T(w-1,s) is the relation
𝔖w-1=xs𝔖u-1+∑v∈Φ𝔖v-1.
We may notice directly one corollary of (4.29). Let
𝔖w(1)=𝔖w(1,1,…)
be the number of monomials in 𝔖w, each counted with its multiplicity. (By (4.17),
𝔖w is a positive sum of monomials.) If T(w,r)
is a transition, we have
𝔖w(1)=𝔖u(1)+∑v∈Φ𝔖v(1)
and also, by (4.29)
𝔖w-1(1)=𝔖u-1(1)+∑v∈Φ𝔖v-1(1).
From these two relations it follows, by induction on ℓ(w) and on the integer
𝔖w(1), that
(4.30)𝔖w(1)=𝔖w-1(1)
or in other words that 𝔖w and
𝔖w-1 each contain the same number of monomials. So if Kohnert's algorithm
(4.20) is true, we should have
CardK(D(w))=CardK(D(w-1)).
Doubtless the combinatorialists will seek a "bijective" proof of this fact.
Let T(w,r) be a transition and let
v∈Φ(w,r).
Consider again the graphs of w and v:
G(w)G(v)
Let m,n,p denote respectively the number of points of
G(w) (or equivalently G(v))
in the open regions of M,N,P. (The regions marked
with a zero contain no graph points.) Then we have
and ck(v)=ck(w)
if k≠i,r. In particular,
cr(w)>cr(v)
for all v∈Φ(w,r).
Proof.
ci(w) is the number of positive integers
k>i such that w(k)<w(i),
hence is equal to m+n. Similarly for the other assertions.
□
Suppose first that m=0, i.e (by (4.27)) that v is the leader of
Φ. Then from (4.31) we have ci(w)=cr(v)
and cr(w)=ci(v).
Hence in this case c(v)=tirc(w)
and therefore λ(v)=λ(w).
If on the other hand m>0, there are two possibilities:
either
ci(v)>ci(w)≥cr(w)>cr(v),
or
ci(v)>cr(w)>ci(w)>cr(v).
In both cases it follows that λ(v) is of the form
Raλ(w), where
R is a raising operator and a≥1. Hence
λ(v)>λ(w)
(for the dominance partial ordering on partitions), and we have proved
(4.32)
If T(w,r) is a transition, we have
λ(v)≥λ(w)
for all v∈Φ(w,r), with equality if and
only if v is the leader of Φ.
Recall (1.26) that for any permutation w we have
λ(w)′≥λ(w-1).
Hence for v∈Φ(w,r) we have
(4.33)λ(w)′≥(*)λ(v)′≥λ(v-1)≥(*)λ(w-1)
by (4.29) and (4.32). Moreover, at least one of the inequalities (*) is strict unless v
is the leader of Φ(w,r) and
v-1 is the leader Φ(w-1,s)
(in the notation of (4.29)). In the notation of the diagram preceding (4.27) this means that
(A′∪B′∪C′)∩G(w)=∅
and hence, as in the proof of (4.26), that CardΦ≤1.
(4.34)
If T(w,r) is a transition with w vexillary, then
Φ(w,r) is either empty or consists of one
vexillary permutation.
Proof.
Suppose that Φ is not empty, and let
v∈Φ. By (1.27) we have
λ(w)′=λ(w-1),
and hence all the inequalities in (4.33) are equalities. Thus v is vexillary, and by the remarks above it is the only member of
Φ.
□
(4.35)
Let T(w,r) be a transition with
r>d0(w). Then
d0(v)≥d0(w)
for all v∈Φ(w,r).
Proof.
As before, let v=utir with
i<r, and let
d0(w)=d. We have to show that
(*)c1(v)≤…≤cd(v).
We distinguish three cases:
(a)
i>d, so that d≤i-1
and therefore ck(v)=ck(w)
for 1≤k≤d.
(b)
i=d. In this case we have
ck(v)=ck(w)
for 1≤k≤d-1, and
cd-1(v)=cd-1(w)≤cd(w)<cd(v)
by (4.31), so that
cd-1(v)<cd(v).
(c)
i>d. Since d<r we have
i+1<r and
ci(w)≤ci+1(w),
hence w(i+1)>w(i).
The diagram on p. 58 shows that w(i+1)>w(j),
or equivalently v(i+1)>v(i),
so that ci(v)≤ci+1(v).
Hence
ci-1(v)=ci-1(w)≤ci(w)<ci(v)≤ci+1(v)
and therefore
ci-1(v)<ci(v)≤ci+1(v).
Since the sequences (c1(v),…,cd(v))
and (c1(w),…,cd(w))
differ only in the ith place, we have
c1(v)≤…≤cd(v)
as required.
□
The maximal transition for w is
T(w,d1(w)).
Let us temporarily write w→v to mean that
v∈Φ(w,d1(w)).
(4.36)
Suppose that
w=w0→w1→…→wp
is a chain of maximal transitions in which none of the wi is Grassmannian. Then
p<(d1(w)-d0(w))ℓ(w).
Proof.
For any permutation v, let
e(v)=d1(v)-d0(v)≥0.
Also let f(v) denote the last nonzero term in the sequence
c(v), i.e.
f(v)=cd1(v)(v).
Recall that v is Grassmannian if and only if it has only one descent, that is to say if and only if
e(v)=0.
From (4.35) we have
d0(wk)≥d0(wk-1)
for 1≤k≤p, and from (4.31) we have
(1)cr(wk)<cr(wk-1)
where r=d1(wk-1).
Hence d1(wk)≤d1(wk-1)
and therefore
e(wk)≤e(wk-1).
Moreover, if e(wk)=e(wk-1)
we must have d1(wk)=d1(wk-1) and hence by (1)
f(wk)<f(wk-1).
It follows that the p+1 points
(xk,yk)=(e(wk),f(wk))
are all distinct. Since they all satisfy 1≤wk≤e(w)
and 1≤yk≤ℓ(w),
we have p+1≤e(w)ℓ(w),
as required.
□
The rooted tree of a permutation
In what follows we shall when necessary replace a permutation w by 1×w,
in order to ensure that at each stage the set Φ(w,r)
is not empty (4.28). Observe that this replacement does not change the bound
(d1(w)-d0(w))ℓ(w)
in (4.36).
The rooted treeTw of a permutation w defined as follows:
(i)
if w is vexillary, then
Tw={w};
(ii)
if w is not vexillary, take the maximal transition for w:(*)𝔖w=xr𝔖u+∑v∈Φ𝔖v
where r=d1(w). (If
Φ is empty, replace w by 1×w
as explained above.) To obtain Tw, join w by an edge to each
v∈Φ, and attach to each
v∈Φ its tree Tv.
By (4.36), Tw is a finite tree, and by construction all its endpoints are vexillary permutations of length
ℓ(w). It follows from (4.28) that
v↦1×v is a bijection of Tw onto
T1×w. Thus Tw
depends (up to isomorphism) only on the diagonal equivalence class (Chapter I) of the permutation w.
Recall that ρm:P∞→Pm
is the homomorphism defined by ρm(xi)=xi
if 1≤i≤m, and
ρm(xi)=0 if
i>m.
(4.37)
Let V be the set of endpoints of Tw. Then if
m≤d0(w) we have
ρm(𝔖w)=∑v∈Vsλ(v)(Xm).
Proof.
If w is vexillary we have ρm(𝔖w)=sλ(w)(Xm)
by (4.4), since ϕ1(w)=d0(w)≥m.
If w is not vexillary, it follows from the maximal transition (*) above that
ρm(𝔖w)=∑v∈Φρm(𝔖v)
since r=d1(w)>d0(w)≥m.
The result now follows by induction on Card(Tw).
□
Multiplication of Schur functions
Let μ,ν be partitions and let
u∈Sn,u′∈Sp be Grassmannian permutations of shapes
μ,ν respectively. Let
w=u×u′∈Sn+p,
so that by (4.6) and (4.8)
𝔖w=𝔖u·𝔖1n×u′=sμ(Xr)sν(Xs)
where r=d0(u) and
s=n+d0(u′).
Hence if m≤r we have
ρm(𝔖w)=sμ(Xm)sν(Xm)
and so by (4.37)
sμ(Xm)sν(Xm)=∑v∈Vsλ(v)(Xm)
where V is the set of endpoints of the tree Tw. Here the integer
m can be arbitrarily large, because we can replace w by
1k×w for any positive integer k. Consequently we have
(4.38)sμsν=∑v∈Vsλ(v)
where V is the set of endpoints of the tree Tu×u′,
and u (resp. u′) is Grassmannian of shape
μ (resp. ν).
The same argument evidently applies to the product of any number of Schur functions. If
μ(1),…,μ(k)
are partitions, let ui∈Sni be a Grassmannian permutation of
shape μ(i), for each
i=1,…,k (so that
ni≥ℓ(μ(i))+ℓ(μ(i)′))
and let w=u1×…×uk. Then
(4.38')sμ(1)…sμ(k)=∑v∈Vsλ(v)
where V is the set of endpoints of the tree Tw.
In particular, suppose that each μ(i) is one-part partition, say
μ(i)=(μi),
so that the left-hand side of (4.38') becomes hμ1hμ2…=hμ.
Correspondingly, each ui is a cycle of length μi+1,
namely ui=(μi+1,1,2,…,μi).
Now [Mac1979, Ch.I, §6] the coefficient of a Schur function sλ in hμ
is the Kostka number Kλμ. Hence we have
(4.39)
Kλμ is the number of endpoints of shape λ
in the tree of w=u1×u2×….