This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
a = p(z) for every polynomial, then z z + such that 1-1= HI.!) , where Hq> ! def P_ q> f , f e. H~, . .
0
So
P
a
=
+ (z-a)q ; hence
p(a)
Since
k
0 ,
~
ran(S-a)
Conversely,
a
k
°P *
is a bpe, then
a
0p(s*)
E
k,
is any polynomial, then there is a polynomial
p(a)
a bpe with
Thus,
then there is a corresponding eigenvector
Thus if
If
for every
extends to a bounded linear functional on
polynomial
k
J jpj2d~
i f there
a
Note that if there is a bpe at a
Hence there is a vector
if
~
jp(a)j2
C2
ran(S-a)
J..
~
a-
p(a)
0 .
~
ker(S-a)
+ a
Thus
is
defines a bijec-
*
is a nontrivial invariant subspace for
~
and so
(0)
S
Much work
has been done on bpe's and it seemed a tractable approach to the invariant subspace problem.
However it has now been shown that
every subnormal operator has an invariant subspace, but it remains unknown whether every non-normal cyclic subnormal operator has a bpe.
(It is worth noting that Fernstrom [13] has given an example
of a rationally cyclic non-normal subnormal operator that has no bpe's.) By the results contained in §2 of this paper, it suffices in proving the invariant subspace theorem to assume (3.1)
the identity map
p
~
p
on polynomials extends to an
isometry and a weak-star homeomorphism of Note that if p(a)
is a bpe and
for every polynomial
implies that on
a
poo(~)
~ ~ f ~kad~
k
a
€
Hoo(ID)
p2(~)
onto
poo(~)
such that
p, then the fact that
L2(~) ~ Ll(~)
is a weak-star continuous homomorphism
Thus by Proposition 2.6,
taj < 1 •
J. B.CONWAY
30
10.1
Now assume that that
p
+
p(o.)
poo(~).
on
= J pko.d~
k
L2(~)
€
Again Proposition 2.6 implies
extends to a weak-star continuous homomorphism
Therefore there is a
p(o.)
a not known.
< 1.
k
for every p01ynOmia1
L1(~)
in o.
p
, then we would have a bpe.
such that
If we also knew that Unfortunately, this is
Nevertheless the invariant subspace problem can be
solved by proving that this
k
has another property and that
a
this suffices to prove that subnormal operators have invariant subspaces. 3.2
THEUREM.
If
S
~
and
(3.1) and if, in addition,
S
trivial reducing subspaces and ran(S-\)
satisfies: (ii)
f
and
g
in
p2(~)
(i)
S
for every
is not closed; then for each
are functions
~
are as above with
a
in
satisfying has no non-
\
in
o(S) n ID
o(S) , there
such that
p(o.) for every polynomial Proof.
p.
SO, p. 415.
3.3 THEOREM (Scott Brown). Every subnormal operator on a space of dimension at least 2 has a non-trivial invariant subspace. Proof. Let S € B(H) and assume S is subnormal. If dim H < 00 ,S is normal and has invariant subspaces. So assume dim H = 0 0 . As was said before, S may be assumed cyclic. So assume that S and ~ are as in the beginning of this section. Also it suffices to assume that ~ .satisfies (3.1). If S has a non-trivial reducing space, then we are done; so assume that S satisfies (i) in Theorem 3.2. If there is a \ in o(S) such that ran(S-\) is closed, then ran(S-\) is an invariant subspace. Now an elementary argument shows that ker(S-\) is a reducing space for S. Since we have assumed (i), ker(S-\) = (0) • Since \ € o(S) , this forces ran(S-\) to be proper and we are done. Thus it suffices to assume that (ii) in Theorem 3.2 holds. Now by (3.1), then
o(S)
-=-
dID
be an isometry.
IISII
and
S
=
1 ; so
o(S)
-=-
dD.
If
D n o(S)
has invariant subspaces since it must
So assume there is a point
a
in
ID n o(S)
o
A SURVEY OF SOME RESULTS ON SUBNORMAL OPERATORS
and let
f, g
span of
{(S-u)nf
g ~
M.
be as in Theorem 3.2. n ~ I} , then
M~ H
Hence
M=
If
trivial.
M=
If
M
the closed linear
is invariant for
If
M~
Sand
ker(S-u) ~ (0)
, then
(0)
this violates (i) in Theorem 3.2.
31
(0)
M
, then
and
is non-
•
One could ask whether there are any subnormal operators satisfying the hypothesis of Theorem 3.2. In fact, getting S and ~ that satisfies (3.1) and such that S is irreducible is rather easy. The unilateral shift and the Bergman shift pop into mind. Whether S can also be chosen with ran(S-A) not closed for every A in o(S) is almost a rephrasing of the bpe problem. Call a subnormal operator pure if there is nu reducing subspace on which it is normal. have no point spectrum.
So i f
point spectrum (that is, inHII(S-A)hll ran(S-A)
If r
Ilhll = l}
is not closed.
restated as
an
:
u
0
>
o(S) =
Now a pure subnormal operator can
A
E
0 0
ap
o ),
(S) denotes the approximate ap (S) means that
(S) if and only if ap Thus condition (ii) in (3.2) can be
(S). ap is a bpe, say that
then
A
E
0
0
u
such that each point in
is an analytic
~
if there is
{z
Iz-ul < r} = B(u;r) is 2 a bpe and i f for each z in B(u;r) k E P (~) such that z
.
by definition,
3.4
abpe(~)
THEOREM (T.T. Trent).
Proof.
is open. abpe(~)
o (S)
\0
ap
(S)
.
See SO, p. 172.
So Theorem 3.4 says that if there were an S satisfying the conditions of Theorem 3.2, then there would be a measure having no analytic bpe's and conversely. This does not say that ~ has no bpe's but it is close. In fact all the cyclic subnormal operators that are known have very thin approximate point spectrum. The following is unknown.
32
J.B.CONNVAY
3.5 PROBLEM. Does there exist a pure cyclic subnormal operator whose approximate point spectrum has non-empty interior? Is there one whose approximate point spectrum has positive area? After we have that subnormal operators have invariant subspaces, then what? Bob Olin and Jim Thomson [r8] have shown that subnormal operators are reflexive. That is, if S is subnormal on Hand T E B(H) such that TM c M whenever M is invariant for S , then T belongs to the weakly closed algebra generated by S. One series of questions concerns the structure of Lat S , the lattice of invariant subspaces of S. For example, the following is unknown. 3.6 PROBLEM. If S is a subnormal operator and Boolean algebra, must S be normal?
Lat S
is a
Another way of asking the question in (3.6) is to ask whether the fact that Lat S is order isomorphic to the lattice of invariant subspaces of a cyclic self-adjoint operator implies that S is normal. §4.
UNITARY EQUIVALENCE, SIMILARITY, AND QUASISIMILARITY
Sl , S2 are operators on HI " H2 , then Sl is guasisimilar to S2 (Sl ~ S2) if there are operators X .. : H. ~ H. 1 1J J (i , j = 1,2) having dense range and trivial kernel and such If
that
X.. S. ~
J
operators.)
SiXij
(Take
XII
and
X22 to be the identity It was shown by Clary [5] that quasisimilar subnormal
(in fact, hyponormal) operators have equal spectrum.
Clary also
managed to characterize the subnormal operators that are quasisimilar to the unilateral shift as well as those that are similar to the unilateral shift.
(See SO, pp. 450-451.)
Quasisimilarity is a rather weak equivalence relation. This means that the equivalence classes are large and so it is not as difficult for two operators to be in the same equivalence class as it is for two operators to be similar or unitarily equivalent. On the other hand it is somewhat difficult to show that quasisimilar operators share certain properties. Thus most of the work on quasisimilarity of subnormal operators has involved efforts to find what properties are preserved under the equivalence.
33
A SURVEY OF SOME RESULTS ON SUBNORMAL OPERATORS
If
51' 52
are subnormal operators with minimal normal
N1 ,N 2 having svsm's ~1' ~2 ' and if 51 ~ 52 ' then I showed in [6] that the identity map on polynomials extends
extensions
poo(~l)
to an isometry of
poo(~2) •
onto
In light of Corollary
2.2, this says that quasisimi1ar subnormal operators generate naturally isometric weakly closed algebras. Marc Raphael has shown that cyclic subnormal operators that are quasisimi1ar have the same e3sentia1 spectrum [19] and have naturally isometric commutants [20]. If the subnormal operators are not cyclic, the equality of the essential spectrum remains an open question. 4.1 PROBLEM. Do quasisimi1ar subnormal operators have equal essential spectrum? My suspicion is that the answer to (4.1) is no, but that it is probably true if the subnormal operators have compact se1fcommutators. In general, quasisimi1ar subnormal operators do not have isometric commutants. In fact, similar subnormal operators may not have isometric commutants (though they clearly have isomorphic commutants). Let's look at an example. Consider on
H2
H2
and let
and
2 a
L (ID) : let
M
=
M = {f $ fl6 : is closed.
be multiplication by
B be multiplication by
K = L 2 (aID) $ L 2 (D) Put
5
z
0e
{re 1
:
on
L2(ID).
Let
a
0 ~ r < 1 , lei <~} •
It is not difficult to see that
(For the details of the proofs of the statements
made here the reader can consult 50, pp. 276-281 or [9].) 2 a
H
M+ [O$L (ID)]
50
K
L2 (aID) $ L2 (D)
Then
M
Let
$ M
z z mne(5 1 ) by
R(f -1
R5 2R onto
kernel
$
=
2
H
z
H
-
{f$fI6+g: fEH
51 = M $ M IH. z z
If
P
Let 2 a
,gEL (ID)}c
H is closed and invariant for 50
51
Let 5 = 5 e B Mz e M z 2 g) f e (f 16 + g). Then 51.
2
is subnormal; also R: H2 $ L2 (ID) ~
Define
a
is invertible and is the orthogonal projection of H2 R
-1
e L2(ID) a
,then RPR is the idempotent with range M and 2 0 $ L (ID). Thus IIRPR- 1 11 > 1 = IIpll In fact any a
isomorphism of
{5}' 2
onto
{5 }' 1
would have to take
P
onto
H
J.B.CONWAY
34
this idempotent since they are the only non-trivial idempotents in the respective commutants.
Thus no isomorphism could be an
isometry. For cyclic subnormal operators a result of Yoshino (SO, p. 147) implies that every operator in its commutant must be subnormal. Since the norm of a subnormal operator equals its spectral radius, quasisimilar subnormal operators have the same norm. This is one of the extra facts about cyclic subnormal operators that is used in Raphael's results. 4.2 PROBLEM. Do quasisimilar subnormal operators have naturally isomorphic Co~~ut3nts? A problem that no one seems to have yet considered is the following. 4.3 PROBLEM. Do quasisimilar subnormal operators have order isomorphic lattices of invariant subspaces? In his thesis Clary answered this question in the affirmative if one of the operators is the unilateral shift of multiplicity one. Now for some similarity. 4.4
PROPOSITION.
then
If
is similar to
Sl
and
Sl S2
S2
i f and only i f there are comEactll
sUEEorted measures
lJ l
and
constants
c2
such that:
cl
lJ 2
on
a:
and there are Eositive
is unitarily equivalent to multiEliS. J cation by z on P (lJ.) ; J (b) For every Eol;tnomial p , (a)
Proof.
For
and
are cyclic subnormal °Eerators
j
1,2 2
Mimic the proof (4.14) on p. 223 of SO.
The cognizant reader will see that the condition (b) is reminiscent of the definition of a Carleson measure.
Indeed
Clary has shown that a subnormal operator is similar to the unilateral shift if and only if S is unitarily equivalent to 2 multiplication hy z on P (lJ) , where lJ is supported on
A SURVEY OF SOME RESULTS ON SUBNORMAL OPERATORS
ciID, III
aD
=
arc length measure, and
(see SO, p. 451).
3S
III ID
is a Carleson measure
Hastings [14] has improved these results.
This is essentially the totality of work cn similarity for subnormal operators. For unitary equivalence, nothing is known other than the" next result.
4.5 i
=
PROPOSITION.
1,2 , then
If
Sl
Si
= multiplication by
z
is unitarily equivalent to
on S2
2
P (Il i )
,
if and only
are mutually absolutely continuous and there is
if in
an
See SO, p. 217.
Proof.
In many ways this result is useless because i t is so difficult to apply.
Our ignorance of the spaces
p2 (11)
is so abysmal p2(1l)
that we can't tell what the moduli of functions in like.
look
Until this function theory is developed, there is little
hope of truly understanding subnormal operators, let alone their unitary equivalence classes. Nevertheless there are several problems that are feasible even in this state of ignorance. Let's see to what extent the general question of unitary equivalence can be reduced to the question for cyclic operators. If
of also
S
E
B(H)
n:?: O}
{Snh
T E B(K)
for each
and
h
in
H,
h E
and let
say that
H
Sh
let
=
Sand
there is a
and, conversely, for each
k
So
SIH h .
Sh
If
is cyclic.
Tare locally equivalent if k
in
the closed linear span
Hh
in
K
K
such that
there is an
h
Sh ~ Tk in
H
such
It isn't hard (if you use the spectral multiplicity theory for normal operators) to show that two normal operators are locally equivalent if and only if they have the Same svsm.
Thus we can never hope to go from local equivalence
to global equivalence in the presence of normality. 4.6 PROBLEM. If two irreducible subnormal operators are locally equivalent, must they be unitarily equivalent? As was mentioned before, local equivalence for normal operators fails to count multiplicities. What if there is no
J.B.CONWAY
36
multiplicity to account for? Specificially, what if the normal operator has infinite multiplicity? 4.7 PROPOSITION. If Nand M are normal operators such that N ~ N $ N , M ~ M $ M , and N and M are locally equivalent, then N ~ M • Proof.
Since
M and
valent svsm's. for both N
are locally equivalent, they have equi-
N.
By multiplicity theory,
has uniform infinite multiplicity. N
M~ N
~
\1
4.8 ~
multiplication by
~
N
\1
$ ... ;
PROBLEM. S
must
~
If
hence
T
z
N~M
Sand
S , T _ T $ T
Sand
~
Thus we may take the same measure
M and
where
S
N
,~nd
T
N
N
~
~
N
as a svsm N
~
~
N
$ N
\1
implies $ ... ,
Similarly,
on
•
are subnormal operators such that Sand
T
are locally equivalent,
be unitarily equivalent?
NOTES 1. Preparation of this paper was partially supported by National Science Foundation Grant MCS 81-21201.
2. The second question raised in (3.5) has an answer - yes. If S is the Toeplitz operator whose symbol is the Riemann map of D onto the inside of a simple closed curve having positive area, then S is the relevant operator. 3. R.F. Olin, using some results from his recent work with J.E. Thomson, has shown that this question has an affirmative answer. REFERENCES M.B. Abrahamse and R.G. Douglas, "A class of subnormal operators related to multiply connected domains," Adv. Math., 19(1976), pp. 106-148. 1.
2. S. Axler and J. Shapiro, "Putnam's theorem, Alexander's spectral area estimate, and VMO," (preprint). 3. J. Chaumat, "Adherence faible ~toile d'alg~bres de fractions rationelles," Ann. Inst. Fourier 24(1974), pp. 93-120. 4. J.A. Gjrna and A. Matheson, polynomials," (preprint).
"Approximation in the mean by
A SURVEY OF SOME RESULTS ON SUBNORMAL OPERATORS
37
5. W.S. Clary, "Equality of spectra of quasisimilar hyponormal operators," Proc. Amer. Math. Soc., 53(1975), pp. 88-90. 6. Ill.
J.B. Conway, "On quasisimilarity for subnormal operators," l. Math., 24(1980), pp. 689-702.
7. J.B. Conway, Subnormal Operators, Research Notes in Math., vol. 51, Pitman Publ., London (1981). 8. J.B. Conway and R.F. Olin, "A functional calculus for subnormal operators, II," Memoirs Amer. Math. Soc., vol. 184(1977). 9. J.B Conway and C.R. Putnam, "An irreducible subnormal operator with infinite multiplicities," l. Operator Theory (to appear). 10. C.C. Cowen, "Subnormality of the Cesaro operator and a semigroup of composition operators," Ind. Univ.Math.J. (1984). 11.
C.C. Cowen and J. Long,
"Some subnormal Toeplitz operators,"
J. Reine Angew.Math. 351 (1984) pp216-220. 12. J.J. Dudziak, "Spectral mapping theorems for subnormal operators," l. Xunctional Anal., (to appear). 13. C. Fernstrom, "Bounded point evaluations and approximation in LP by analytic functions," Spaces of Analytic Functions, Springer-Verlag Lecture Notes, vol. 512(1976), pp. 65-68. 14. W.W. Hastings, "Subnormal operators quasisimilar to an isometry," Trans. Amer. Math. Soc. 256(1979), pp. 145-161. 15. G.E. Keough, "Subnormal operators, Toeplitz operators, and spectral inclusion theorems," Trans. Amer. Math. Soc. 263(1981), pp. l25-l35. 16. T.L. Kriete and D. Trutt, "The Cesaro operator in subnormal," Amer. l. Math. 93(1971), pp. 2l5-225.
£,2
is
17. A. Lubin, "Weighted shifts and products of subnormal operators," Indiana Univ. Math. l., 26(1977), pp. 839-845. 18. R.F. Olin and J.E. Thomson, "Algebras of subnormal operators," l. Functional Anal., 37(1980), pp. 271-30l. 19. M. Raphael, "Quasisimilarity and essential spectra for subnormal operators," Indiana Univ. Math. l., 31(1982), pp. 243-246. 20. M. Raphael, "Commutants of quasisimilar subnormal operators," (preprint). 21. D. Sarason, "Weak-star density of polynomials," J. Reine Angew. Math., 252(1972), pp. 1-15.
OPTIMIZATION, ENGINEERING, AND A MORE GENERAL CORONA TREOREM
J. William Relton University of California, San Diego La Jolla, California 92093, USA
These talks concern a certain optimization problem over the space of functions on the unit circle 1]: It is a general problem which subsumes the classical problem of finding the distance of a given function g to Roo as a very special case. The qualitative theory has advanced rapidly over the last few years to the point where it now strictly extends the main result of the classical Roo distance problem. The talk describes these new results, a more general Corona theory which is a by-product of them, and some of the substantial engineering motivation which underlies the optimization problem. The optimization problem we study goes as follows. Let r (e i8 ,z) be a function on x ~N and let E be a subset of 1f(~N) the vector valued Loo functions on lr The basic problem is:
lr
(OPT) Find 10, find optimizing fa's, say what you can about them even if explicitly .is imfossible. For example, fix a function g and set f(e 1 ,z) = Ig(eI8 )_z I . Then (OPT) with E = Roo is the classical problem of finding the distance of g to Roo. There are both qualitative and for very special cases quantitative results. This talk focuses on the qualitative; since other talks, for example, Sarason's and Young's deal with quantitative issues. We restrict E to be vector valued Roo or to be A = C n ROO. Our main qualitative result for scalar valued functions, namely, when E:= A(~l), is co~rutation
39
s. C
Power (ed.), Operators and Function Theory, 39-40.
40
J. W.HELTON
Theorem 1. If r ·9
defined by a(e l
)
is smooth and a function fo in A
ar = --
az
·9
·9
(e l ,fo(e l
))
makes the function a
non-vanishing, then fo is a local minimum
for (OPT) if and only if (i) r(e i9 ,fo(e i9 )) is constant (independent of 0). (ii) The winding number of a about 0 is positive. This actually provides a practical test which a designer could use to tell if he has a local optimum. Next we turn to A(~N) __ the vector valued case. We show that
(1)
(2) (3)
When all functions are rational, Theorem 1 generalizes perfectly. Again we get two conditions, (i) and (ii), which characterize local minima. Condition (ii) in (1) is not stable under small changes in fo so (1) is not a practical test. To give a stable version of (ii) we need to introduce a measure of the numerical 'conditioning' of (ii). The obvious one is a 'Corona constant' for the functions .~
ar
aj(e IV ) = -
aZj
·9
.~
(e l ,f(e IV ) )
•
Unfortunately, the aj are typically not in Hoc. Thus classical Corona theory does not apply. , (4) We extend the Toeplitz version of the Corona Theorem to the case at hand. The traditional Corona Theorem concerns HOC; ours is a gen(analogous to Takagi's generalization of eralization to Nevanlinna-Pick interpolation from HOC to Hf'). Thus we obtain a natural generalization of this classical result, as well as, obtain a numerically stable test to determine if a given function fo is a (local) optimum for (OPT) with E = A(~N).
Hi
LECTURE FORECAST Lecture 1.
Overview, engineering and the problems it produces. Theorem 1.
Lecture 2.
Survey of results.
Lecture 3.
Proof of the general Hf' Corona Theorem.
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
M.A. Kaashoek Vrije Universiteit Department of Mathematics and Computer Science P.O. Box 7161 1007 MC Amsterdam, The Netherlands In three lectures a survey is given of recent work on factorization and invertibility problems concerning systems and integral operators. In the first lecture the emphasis is on minimal factorization of linear time invariant systems. Wiener-Hopf integral equations with matrix-valued kernels form the main topic of the second lecture. The classical approach to such equations is combined with the method of factorization discussed in the first lecture to obtain explicit formulas for the resolvent kernel and the Fredholm characteristics. Special attention will be paid to equations with analytic symbols and to Wiener-Hopf equations appearing in transport theory. In the third lecture integral operators are viewed as transfer operators of linear time varying systems with well-posed boundary conditions. This approach allows one to study invertibility and factorization of integral operators in terms of inversion and decoupling of systems. The main part of the work reported on is joint work (in different groupings) with H. Bart, I. Gohberg, F. van Schagen and P. Van Dooren. It is a pleasure to thank I. Gohberg for several discussions about the plan for these lectures. I.
FACTORIZATION PROBLEMS
Factorization problems appear in different areas of analysis and its applications. In this first lecture three examples are described, each of which may be viewed as a different version of the problem of minimal factorization. For regular rational matrix functions the solution of the latter problem is given in detail. I.l. Wiener-Hopf integral equations and canonical factorization A finite system of Wiener-Hopf integral equations may be written in vector form as a single equation: 00
(1.1 )
f
o
k(t-s)
S. C. Power (ed.), Operatorrand Function Theory, 41-86. _ .1'10" .... ""_.1..1_' LU'_LJI___ " ' _ _ _ _ _ L~
ft...
.0
o
S t <
00.
42
M. A. KAASHOEK
Here f,g: [0,00) + ~m, and throughout this text the kernel k is assumed to be an mXm matrix function of which the entries are in Ll (~,oo). The given function f E Lm[O,oo), i.e., the components of f are in L [0,00). p
~ E
The problem is to find
o
m
p
L [0,00) such that (1.1) holds true a.e. on p
t < 00. The index p is assumed to be fixed and 1 ~ P ~ 00. The classical method to deal with equation (1.1) works as follows (see [22]). Assume that ~ is a solution of (1.1). Extend ~ and f to the full line by defining them to be zero on the negative part of the real line. Then, by introducing an appropriate function fO E L;[O,oo) of which ~
the support is on t
~
0, equation (1.1) can be rewritten in the form
00 ~(t) -
(1.2)
f
k(t-s)~(s)ds = f(t) + fO(t),
_00
< 00
< t
~
Now apply the Fourier transformation to both sides of (1.2) and leave the part that is given in the right hand side. This yields: (1.3)
Here K(A)
=
f
The function
~+
e itA k(t)dt is the Fourier transform of the kernel k. is the Fourier transform of the function
subindex + is added to indicate the fact that
~
and the
~
is analytic in the + open upper half plane with boundary values up to the real line. Further 00 o iAt f e iAt f(t)dt, F (A) = f e fO (t)dt. F+(A)
o
~
In equation (1.3) the function F+(A) is known and the problem is to find
~+
and F
such that (1.3) holds.
The function weAl
=
I - K(A) appearing in (1.3) is called the
symboL of the equation (1.2). To solve equation (1.3) one looks first of all for a so-called (right) canonicaL factorization of the symbol relative to the real line: (1.4)
_00 < A < 00,
weAl = W (A)W (A), +
where
00
W+(A)
I
m+
f
e
iAt
0 0
f
gl(t)dt, det W+(A)
~
0
(1m A
~
0) ,
iAt
(1m A ~ 0), g2(t)dt, det W_(A) ~ 0 -= with gl and g2 mXm matrix functions of which the entries are in L 1[0,00) W (A) = I
m
+
e
and Ll (-=,0], respectively. If the symbol weAl admits such a canonical
43
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
factorization, then (1.5)
W+O.)-
1
f
00
= 1m +
o
e
iAt 1 0 iAt Y1(t)dt, W_(A)- = 1m + e Y2 (t)dt,
f
~
where Y1 and Y2 are mXm matrix functions of the same type as gl and g2' respectively. Using (1.4), equation (1.3) can be rewritten as ~
(1.6)
Let P be the projection acting on the Fourier transform of Lm(...oo p , 00)_ functions according to the following rule:
~~ eiA~(t)dt)
=
I
eiAth(t)dt.
Apply this projection to (1.6). Then, omitting the variable A, one gets if.. ~+
= W+-l Pw_-l F+,
' wh'1C h'1S t h e Four1er trans f orm 0 f t h e so 1 u t'10n 0 f
It remains to apply the inverse Fourier transform to
~+
(1 • 1)•
to obtain the
solution. The final result is the following theorem (see [22J).
For each f E Lm[O,OO) equation (1.1) has a unique solution p ~ E Lm[O,OO) if and only if the symbol of the equation admits a canonical p factorization relative to the real line, and in that case the solution ~ of (1.1) is given by
THEOREM 1.1.
00
~(t)
= f(t)
+
f
o
o
y(t,s)f(s)ds,
~
t <
00,
where s
(1. 7)
yet,s)
y 1 (t-s) +
f
Y2 (t-s) +
f
{
o
Yl (t-r)Y2(r-s)dr,
t > s,
Yl (t-r)Y2(r-s)dr,
t < s,
t
o
the matrix functions Yl and Y2 being
dete~ined
by (1.5).
For the factorization method to be effective it is important to have conditions that guarantee the existence of a canonical factorization and to have explicit formulas for its factors. For matrix symbols this is a non-trivial problem. In the scalar case the function weAl admits a canonical factorization if and only if the curve parametrized by A ~ W(A), _00 < A < 00, does not go through zero and has winding number zero around zero. Furthermore, in that case exp(P log W(A»,
..00
<
A<
00.
Because of lack of commutativity, the matrix case is much more complicated.
44
M.A.KAASHOEK
If the determinant of the symbol WeAl does not vanish on the real line (which is a necessary condition for canonical factorization), then WeAl admits a (right) Wiener-Hopf factorization:
(l.8)
W (A)
weAl
..00
< A<
00,
where W (A) and W (A) have the same properties as in (1.4). The integers
-
+
Kl ~ •.. ~ Km in (1.8) are the so-called
(right) factorization indices,
which are uniquely determined by w(A). From a Wiener-Hopf factorization one can read off the Fredholm properties of the equation (1.1). The problem is that, in general, one cannot tell in advance when the factorization indices will be zero and the corresponding factorization will be canonical. Also, in general, there are no explicit formulas like in the scalar case. To find conditions that guarantee canonical factorization and to construct explicit formulas for the factors in such a factorization are some of the problems we shall deal with. But, first, let us see two other examples in which factorization plays a decisive role. 1.2.
Characteristic operator functions and their divisors
Let A be a bounded linear operator acting on a separable Hilbert space In a number of instances the problem to find the invariant subspaces of A turns out to be a factorization problem. Let us illustrate this for the simplest case when A is dissipative. So we require the imaginary
H.
part A.
l.m
(2.1 )
1 (A-A) * to be a non-negative operator, and hence 2i
Aim
=
1 (A-A) * 2i
= C*C,
where C is an operator acting from H into some auxiliarly Hilbert space G, which may be different from H. The operator function (2.2) which has to be considered in the open lower half plane and in a neighbourhood of infinity on the Riemann sphere, is called a (Liv~ic) cr~racteristic operator function. In (2.2) the symbol IG denotes the identity operator on the space G. If GO = 1m C, then relative to the decomposition G = GO can be written as:
$
G~ the characteristic operator function weAl
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
45
where WO(A) is determined uniquely by A up to a unitary equivalence. Recall that an operator A is completely nonselfadjoint (shortly: c.n.s.) if there is no non-zero subspace invariant under A and A* on which A acts as a selfadjoint operator. If two c.n.s. dissipative operators Al and A2 have the same characteristic operator function, -1
then Al = UA2 U for some unitary operator U. Thus for c.n.s. dissipative operators the characteristic operator function is a unitary invariant. Let A: H + H be a c.n.s. dissipative operator, and let (2.2) be its characteristic operator function. Assume that M is an A-invariant subspace of H. Write A and C as operator matrices relative to the orthogonal decomposition
A
H=
M e Mi:
C
Then Al and A2 are c.n.s. dissipative operators, the characteristic operator function weAl of A factors as follows: (2.3)
weAl
=
. (A-A ) -1 C* ], EIG + 2iC l (A-AI) -1 C*1 ]EI G + 21C 2 2 2
and the factors are characteristic operator functions of Al and A2 , respectively. Thus an invariant subspace of A leads to a factorization of weAl with factors that are again characteristic operator functions of c.n.s. dissipative operators. What about the converse? Assume that the characteristic operator function weAl of the c.n.s. dissipative operator A admits a factorization weAl = WI (A)W 2 (A), where for V = 1,2 Wv(A)
=
IG + 2iCv(A-~) -1 Cv*
is the characteristic operator function of the c.n.s. dissipative operator Av. Does it follow that the factorization weAl = WI (A)W 2 (A) may be obtained from an invariant subspace of A in the way described in the previous paragraph? The answer is yes if and only if the dissipative operator
46
M. A. KAASHOEK
is also c.n.s., in which case the factorization W(A) called regular. The theorem is (see [12], Chapter I) that the factorization method of the previous paragraph gives a one-one correspondence between the invariant subspaces of A and the different regular factorizations of its characteristic operator function. Thus to find the invariant subspaces of A one has to find the regular factorizations of its characteristic operator function. I.3.
Intermezzo about linear systems and realization of transfer functions
In state space representation a linear time invariant dynamical system has the following form:
8
{x
Ax(t) + Bu(t),
(t)
Cx(t) + Du(t).
y(t)
The main coefficient A is a linear operator acting on a finite dimensional linear space X, which is called the state space of the system 8. The input u(t) is an ~-vector function and the input
coefficient B is a linear operator from ~ into X. The output y(t) is an ~l-vector function and th~ output coefficient is a linear operator from X into e~ternal
~l.
The linear operator 0:
~
+
~l
is called the
coefficient of 8. To simplify the notation we shall represent
the system 8 by its four coefficients and write 8 = (A,B,C,O). Assume the system 8 to be at rest at time t = O. Then one can apply the Laplace transform to rewrite 8 in the form:
'" Cx(A) '"
'"
Ax (A) + Bu (A) ,
'"
+ Du(A) ,
and hence (3.1)
'" '" y(A) = (0 + C(A-A) -1 B)U(A).
Formula (3.1) describes the input/output behaviour of the system 8 in the so-called frequency domain. The function Wa(A) = 0 + C(A-A)-l B, which is called the transfer fUnction of a, is an lXm rational matrix function, which is ana1ytic at infinity. (Throughout this paper we identify an pXq matrix T with the linear operator from ~q into ~ given by the canonical action of T with respect to the standard bases of ~q and ~.) According to a basic theorem in systems theory any rational matrix function which is analytic at infinity appears as the transfer
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
47
function of a time invariant system 8, and given W(A) there are various ways to construct such a 8. The system 8 = (A,B,C,D) is called a reaZization of the rational matrix function W(A) if (3.2)
D + C(A-A)
-1
B.
We shall use the term realization also for the right hand side of (3.2). Realizations are not unique. In fact, if 8 0 = (AO,BO,CO,D O) is a realization of W(A), then the system 8 = (A;B,C,D), where D = DO
* (3.3)
A
o o
* :].
B~
J
C
[0
Co
*J,
is again a realization of W(A). In (3.3) the stars denote unspecified entries and the partioning of the coefficients A, B, C is made relative to a decomposition, X = Xl ~ Xo ~ X2 ' of the state spacp. of 8. If two systems 86 and 8 are related in this way, then 8 is called a diZation of 8 0 and 8 0 is a reduation of 8. A realization 8 of W(A) is called minimaL if among all realizations of W(A) the state space dimension of 8 is as small as possible. Minimal realizations are unique up to similarity, that is, if 8 1 = (A l ,B l ,C l ,D l ) and 8 2 = (A 2 ,B 2 ,C 2 ,D 2 ) are minimal realizatio~s of the same transfer function, then Dl = D2 and there exists a unique invertible operator S: Xl + X2 ' Xl and X2 being the state spaces of 8 1 and 8 2 , respectively, such that (3.4) This result is known as the state spaae isomorphism theorem. The operator S in (3.4) is called a system simiLarity. It can be proved that any realization can be reduced to (or, equivalently, is a dilation of) a minimal realization. Hence, two realizations of the same transfer function are dilations of similar (minimal) realizations. A system 8 = (A,B,C,D) is said to be minimaL if 8 is a minimal realization of its transfer function. The minimal systems coincide with the controllable and observable systems, which means that e = (A,B,C,D) is minimal if and only if (3.5)
Ker C n Ker CA n ••• n Ker CAn - 1 = {oJ,
(3.6)
n-l Im B + lm AB + ..• + lm A B = X,
where n is the dimension of the state space X.
48
1.4.
M.A.KAASHOEK
Cascade connections and minimal factorizations
Let 8 1
=
(A 1 ,B 1 ,C 1 ,01) and 8 2
(A 2 ,B 2 ,C 2 ,02) be time-invariant systems.
Assume that the output space of 82 is equal to the input space of 8 1 • The system 8 which one obtains if the output of 8 2 is equal to the input of 8 1 is called the cascade connection of 81 and 82 . We shall also say that 8 is the product of 81 and 82 and we shall write 8 = 8 1 82 • In the time domain the input/output behaviour of 8 = 8 1 82
r
is described by A1 xl (t) + B1C2x 2 (t) + B10 2u(t),
t
Thus 8
x2 (t))
A2x 2 (t) + B2U(t),
y(t)
C1x 1 (t) + °l C2x 2(t) + °1 0 2u (t).
8 18 2 A
(A,B,C,O) , where °
B,c,].
[:'
=
°1°2 and
_ [B,D2] ,
B -
A2
B2
C
[C 1
°l c 2 J ·
From the definition of a cascade connection it is clear that the transfer function of a cascade connection is the product of the transfer functions of the compounding systems, in other words, (4.1)
Thus cascade connections correspond to factorizations of transfer functions. In order to obtain in the simplest way a system with a prescribed input/output behaviour (or, equivalently, with a prescribed transfer function) by a cascade connection of more elementary systems, it is important to know when a cascade connection of two minimal systems is again minimal. This leads to the problem of minimal factorization of rational matrix functions. Let W(A) be an [xm rational matrix function which is analytic at 00 Thus for A sufficiently large W(A) admits the following expansion: W(A)
=
° +
I °1
+
A~
°2 + . . . .
By definition the MacMillan degree o(W) of W is the integer
(4.2)
o(W)
max rank p
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
49
Let H be the block Hankel matrix in the right hand side of (4.2). p
, 1
e=
If
(A,B,C,D) is a realization of W(A), then D, j = 1,2, ... , and thus J
= CA J -
B,
C
CA (4.3)
p-1 • [BAS ... A BJ.
H
P
. p-1 CA
This implies that rank H1
~
rank H2
~
... and the sequence is bounded
by the dimension of the state space X of 8. In particular, if 8 is minimal, then rank H dim X for p sufficiently. Thus the MacMillan p
degree 6(w) is equal to the state space dimension of a minimal realization of W(A). The latter fact also implies that the MacMillan degree has the following sublogarithmic property: (4.4) where W1 (A) and W2 (A) are rational matrix functions that are analytic at infinity. The factorization W(A) = W1 (A)W 2 (A) is called minimal if
6 (W) = 6 (W 1 ) + 6 (W 2 ) .
(4.5)
The aasaade aonneation e = 8 1 8 2 is minimal if and only if 8 1 and 8 2 are minimal systems and the faatorization We(A) = w8 (A)W 8 (A) is a minimal faatorization.
THEOREM 4.1.
1
2
In the definition of minimal factorization we have restricted ourselves to functions that are analytic at infinity, but by applying a suitable Moebius transformation the analyticity at infinity may be dropped and one can extend the definition of minimal factorization to arbitrary rational matrix functions. A rational matrix function W(A) is called regular if W(A) is a square matrix function and det W(A) does not vanish identically. In that case W(A)-1 is again a rational matrix function and, by definition, -1
a zero of W(A) is a pole of W(A) . For two regular rational matrix functions W1 (A) and W2 (A) the factorization (4.6)
is minimal if and only if there is no pole-zero cancellation between the factors. For the precise meaning of the notion of pole-zero cancellation we refer to [3J. The characterization of minimal factorization in terms of pole-zero cancellation applies also to regular rational matrix functions that are not analytic at infinity.
so
M.A.KAASHOEK
for with I.5.
The solution of the problem of minimal factorization
Let weAl be an mXm regular rational matrix function. We shall describe all minimal factorizations of WeAl of which the factors are again regular matrix functions. Without loss of generality (apply a Moebius transformation if necessary) we may assume that D = limA~W(A) exists and is invertible. This allows us to see weAl as the transfer function of a system 8 = (A,B,C,D) of which the external coefficient is invertible. Thus weAl = D + C(A-A)-l B and, because of the invertibility of the external coefficient D, (5.1)
Note that W(A)-l is the transfer function of the inverse of the system 8 which one obtains by reversing the roles of the input and output. x -1 x We write A for the operator A - BD C, but observe that A does not depend on A only, but also on the other coefficients of 8. We shall x
refer to A as the associate main coefficient of 8. Let 8 = (A,B,C,D) be a system with an invertible external x
coefficient, and let A be its associate main coefficient. A projection IT of the state space X of 8 is called a supporting projection of 8 if x
A Ker IT c Ker IT, A
Im IT c Im IT.
THEOREM 5.1. Let weAl be the transfer function of a system 8 = (A,B,C,D) with an invertible external coefficient~ and let IT be a
supporting projection of 8. PUt (5.2)
Wi,IT(A)
D1 + cIT (A_A)-l BD;l ,
(5.3)
Wr, IT(A)
D2 + D1 C(A-A)
-1
-1
(I-IT)B,
where D1 and D2 are invertible matrices such that D (5.4) w1 (A)W 2 (A) and weAl = W3 (A)W4 (A) are called -1 equal if there exists an invertible matrix E such that W1 (A) = W3 (A)E Two factorizationsW(A)
=
and W2 (A) = EW4 (A). With this definition of equality the factorization (5.4) is uniquely determined by IT and does not depend on the particular factorization of D. THEOREM 5.2.
Let weAl be the transfer /Unction of a minimal system 8
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
51
with an invertible external coefficient. Then the map n + Wt,n(A)Wr,n(A)
defines a one-one correspondence between all supporting projections of e and the different minimal factorizations of W(A) with square factors. By allowing different realizations of W(A) one can use Theorem 5.1 to get all factorizations W(A) = W1 (A)W 2 (A) of which the factors are regular rational matrix functions that are analytic at infinity. However, to get all such factorizations one cannot work with one single realization. When one restricts the attention to minimal factorizations, then the situation is different. According to Theorem 5.2 all minimal factorizations of W(A) of which the factors are regular matrix functions may be obtained in the way of Theorem 5.1 from one minimal realization. Moreover, in that case the correspondence between the supporting projections and the minimal factorizations is one-one. 1.6.
Regular factorization as an infinite dimensional analogue of minimal factorization.
Let us return to the Livsic characteristic operator function of a dissipitative operator A: W(A) = I + 2iC(A-A)-l c*.
(6.1)
In this section we view (6.1) as the transfer function of a system with an infinite dimensional state space and an infinite dimensional input/ output space. A system e (A,B,C,I) is called a Liv~ic system if A: H + H, B: G + Hand C: H + G are bounded linear operators acting between Hilbert spaces, I is the identity operator on G and A - A*
(6.2)
= BC,
B
Its transfer function We(A)
= 2iC * . =I
+ C(A-A)
-1
B, which is equal to the
function (6.1), has the following properties: (1) W(A) is an operator function of which the values are bounded linear operators on a Hilbert space G, (2) W(A) is analytic on an open set nw of the Riemann sphere containing the open lower half plane and the point infinity; (3)
W(A)*W(A)
(4)
W(A)*W(A)
(5)
W(oo)
$
I for A
€
I for A
€
nw, nw'
1m A < 0; A real;
= I.
Conversely, any operator function W(A) with the properties (1)-(5) can be realized as the transfer function of a Liv§ic system. A Liv§ic system e = (A,B,C,I) is called minimaZ (cf., (3.5) and (3.6» i f
52
M.A.KAASHOEK 00
n
(6.3)
00
Ker CA j
V
{O},
j=O
1m AjB
H.
j=O
n N. is defined to be the linear hull of U Nj .) j=O J j=O Because of (6.2), it is easily seen that (6.3) is equivalent to the requirement that the main operator A of 6 is completely nonselfadjoint. A Livsic system can be reduced to a minimal Livsic system without changing its transfer operator. Further, for minimal Livsic systems the following version of the state space isomorphism theorem holds true: If two minimal Livsic systems 6 1 = (A 1 ,B 1 ,C 1 ,I) and 6 2 = 00
(Here
V
(A 2 ,B 2 ,C 2 ,I) have the same transfer function, then there exists a unitary operator U such that
The class of Livsic systems is closed under cascade connections. From the definition of a regular £actorization in Section 1.2 it follows that a cascade connection 6 = 6 1 6 2 of two Livsic systems is minimal if and only if 6 1 and 6 2 are minimal and the factorization W6 (A) = W6 (A)W 6 (A) is regular. By comparing this with Theorem 4.1 one 1 2 sees that regular factorization is an infinite dimensional analogue of minimal factorization. Let 6 = (A,B,C,I) be a Livsic system. Since the external operator of 6 is invertible,
x -1 I - C(A-A) B, x
where A
= A - BC is the associate main coefficient. From the first
identity in (6.2) it follows that Ax = A* . Now, let M be an invariant subspace of A, and let IT be the orthogonal projection onto ML. Then A Ker IT c Ker IT,
AX 1m IT c 1m IT,
and thus IT is a supporting projection of 6. The corresponding factorization of W(A) = W6 (A) is precisely the factorization (2.3). The one-one correspondence between the invariant subspaces of a c.n.s. dissipative operator A and the regular divisors of a characteristic operator function of A may now be translated into a one-one correspondence between the orthogonal supporting projections of the minimal Livsic system 6 and the minimal factorizations of its transfer function. The latter statement is the analogue of Theorem 5.2 for Livsic systems. In general, the factors corresponding to an arbitrary (non-orthogonal) supporting projection of a Livsic system are not transfer functions of Livsic systems.
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
I.7.
S3
An application
We return to the problem of canonical factorization discussed in the first section. Assume that the symbol weAl of equation (1.1) is rational. Since the function k(t) is integrable on ~ < t < ~, the symbol weAl is analytic at infinity and limA~ weAl = I. It follows that weAl may be viewed as the transfer function of a minimal system e = (A,B,C,I) of which the external coefficient is the mXm identity matrix. Now assume that weAl admits a canonical factorization relative to the real line: ~
(7.1 )
< A <
~.
Then the functions W (A) and W+(A) are mXm rational matrix functions, W (A) has no poles and zeros in the closed lower halfplane and W+(A) has no poles and zeros in the closed upper half plane. In particular, weAl and weAl
-1
have no poles on the real line. Using the minimality x
of e this implies that A and A have no eigenvalues on the real line. The fact that the poles and zeros of W_(A) are separated from the poles and zeros of W+(A) implies that in (7.1) there is no pole-zero cancellation between the factors. Hence (7.1) is a minimal factorization and there exists a supporting projection IT of e which yields the factorization (7.1) (cf., Theorem 5.2). So x A Im IT c Im IT,
A Ker IT c Ker IT,
and we may assume that the factors in (7.1) are equal to I + cIT (A-A)-l B,
I + C(A-A)-1(I-IT)B. Using the analyticity of W (A)±1 in the lower half plane and the analyticity of W+(A)±1 in
~he
upper half plane, one finds that Ker IT
has to be the spectral subspace of A corresponding to the eigenvalues of A in the upper half plane and Im IT has to be the spectral subspace
x
x
of A corresponding to the eigenvalues of A in the lower half plane. Proceeding in this way, one comes to the following theorem.
Let weAl = I + C(A-A)-1 B be a minimal pealization with state space x~ and let AX = A - Be. In oPdep that WeAl admits a canonical factoPization pelative to the peal line the following two conditions aPe necessaPY and SUfficient: (i) A and AX have no eigenvalues on the peal line and (ii) the state space x admits the THEOREM 7.1.
54
M. A. KAASHOEK
following direct sum decomposition: x
(7.2)
where
X=MEBM, M
is the spectral subspace of A corresponding to the eigenvalues
of A in the upper half plane and MX is the spectral subspace of AX corresponding to the eigenvalues of AX in the lower half plane. Further, in that case, the factorization (7.3)
where IT is the projection of X along M onto MX, is a right canonical factorization of W(A) relative to the real line. X
The spaces M and M X
M = 1m P and M (7.4)
appearing in (7.2) can also be described as
x
= 1m P , where P and P
1 P = ~ lT~
f f
(A-A) -1 dA,
x
x P
are the Riesz projections,
1 = 2lTi
!
+
(A_A X )-l dA ,
+ x
corresponding to the eigenvalues of A and A in the upper half plane. Thus in (7.4) the curve f+ is a curve in the upper half plane around
x
the eigenvalues of A and A in the upper half plane. Condition (i) in Theorem 7.1 guarantees that W(A) has no poles and zeros on the real line. Thus, if (i) holds, then W(A) admits a right Wiener-Hopf factorization relative to the real line (cf., formula (1.8». If, in addition, (7.5)
MOM
x
x ~ {a} and/or M + M ~ X,
and thus (ii) is not satisfied, then this factorization is non-canonical and there are non-zero factorization indices. It turns out that the x
x
non-zero factorization indices may be read off from MOM and M + M • In fact, there are precisely t negative factorization indices K1 , ... ,K t and precisely s positive factorization indices ~l' .•. '~s' where MoM
t = dim
x
x
s = dim M+M +Im B
MOMxliKer C
#{v
K.
J
~i
I
dim
j
1, •.• ,t
#{v
I
i
dim
x M o M o Ker C 0 x M o M o Ker C 0
x M + M + 1m B + x M + M + 1m B +
1, ••• ,s
M+M
x
v-2 o Ker CA V-l o Ker CA + 1m AV- 1 B
V-2 B
+ 1m A
~
~
j},
s-i+l },
55
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
The above formulas are also valid for non-minimal realizations of W(A) x
provided A and A have no real eigenvalues. I.B.
Further remarks and references.
For Section 1.1. we refer to [22J and for Section 1.2 to [12J. The theory of characteristic operator function is a large subject, which plays an important role in the structure theory of Hilbert space operators (see, e.g., the books [12, 23, 37J). The idea to see the Livsic characteristic operator function as a transfer function is already present in [12J. The analogy between regular factorization of characteristic operator functions and minimal factorization of rational matrix functions has been observed in [3J. Polynomial factorization, which is not discussed here, is in the case of monic operator polynomials also a minimal factorization. The one-one correspondence between the supporting subspaces of a monic operator polynomial and its right divisors in the Gohberg-Lancaster-Rodman divisibility theory of operator polynomials (see [24J) may be seen as a variant of Theorem 5.2 (see [3J, also [2J). The state space approach to systems theory goes back to R.E. Kalman and the description of all realizations of a transfer function in terms of dilation, reduction and similarity, given in Section 1.4, is his canonical structure theorem specified for time invariant systems (see [2BJ). For books on systems theory see [27, 29, 35J. The solution of the problem of minimal factorization for regular rational matrix functions is taken from [10J. For the case when the external coefficient D is not invertible a direct analysis of the . minimal factorization problem, avoiding Moebius transformations, is given in [14J. The problem of minimal factorization for sigular rational matrix functions (with a somewhat different notion of minimal factorization) is solved in [3BJ. Theorem 7.1 on canonical factorization is taken from [3J. The problem to construct a non-canonical Wiener-Hopf factorization is not a problem of minimal factorization. In fact, to make such a -1 factorization by starting from a realization W(A) = I + C(A-A) Band by using the geometrical factorization principle of Theorem 5.1 is much more involved. It requires the construction of a special dilation G = (A,B,C,I) of the system e = (A,B,C,I), which permits to obtain the factors in terms of invariant subspaces of A and A - BC. The final formulas are again in terms of the original coefficients and certain auxiliary operators introduced in the construction of the dilation (see [6, BJ). The formulas for the indices given at the end of Section 1.7 also appear in [6, BJ. The description of factorization in terms of invariant subspaces
e
x
of A and A is very convenient for the study of stability properties of the factors under small perturbations and led to interesting questions, both in the finite and infinite dimension setting, concerning stably invariant subspaces (see [3J, Chapter B; [1J, [17], [34]).
M.A.KAASHOEK
S6
II.
WIENER-HOPF INTEGRAL EQUATIONS
In this chapter the solution of the canonical factorization problem obtained in the previous chapter is used to derive invertibility conditions and resolvent kernels for Wiener-Hopf integral equations. First the case of rational symbols is considered. Next the same method is applied to certain Wiener-Hopf integral equations with non-rational symbols arising in transport theory and, more generally, to equations with symbols that are analytic in a strip around the real line. In the last section realization of the symbol is used to make a direct connection between Wiener-Hopf integral equations and time-invariant linear systems. This second approach, which allows one to derive explicit formulas for solutions, Fredholm characteristics, asymptotics, etc., does not use factorization and also applies to convolution equations on a finite interval. II.l.
Invertibility and resolvent formula
Consider the Wiener-Hopf integral equation 00
(1.1)
-f
k(t-s)<j>(s)ds = f (t) ,
0
~
t <
00
0
As before, in (1.1) the function k(t) is an mXm matrix function of which the entries are in Ll (~,oo). We assume that the symbol weAl is rational. We shall combine the canonical factorization theorem of Section 1.7 with the general factorization method outlined in Section 1.1 to solve equation (1.1). The first step is to make a minimal realization of the symbol: 0.2)
Since weAl is continuous on the real line, the minimality of the realization (1.2) implies that A has no ~igenvalues on the real line. x
Next, assume that also A = A-BC has no eigenvalues on the real line (which is equivalent to the requirement that det WeAl ~ 0, ~ < A < 00) • x x Let M (resp., M ) be the spectral subspace of A (resp., A ) x corresponding to the eigenvalues of A (resp., A ) in the upper (resp., x lower) half plane. Assume X = M $ M , where X is the state space of the realization (1.2). Put I + CIT(A-A)-l B,
~
<
A< ~
<
00,
A<
00,
where IT is the projection of X along M onto MX. Then weAl = W_(A)W+(A) is a canonical factorization relative to the real line of the symbol of (1.1) (see Theorem 1.7.1). It follows that (1.1) is uniquely
57
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
solvable in Lm[O,OO)
(see Section 1.1). To find the resolvent kernel we 1 -1 have to compute the inverse Fourier transforms of W (A)- and W+(A) . To do this one first checks that p
-<XJ
< A<
00,
x Next, using the special properties of the spaces M and M , one proves that o x W (A)-l I + f ieiAtce -iA SrBdt, 00
W+(A)-l
=
I -
f
x ieiAtC(I_TI)e-iA t dt .
o Finally, one applies formula (1.1.7) the resolvent kernel.
~o
obtain from these identities
Let WeAl = I + C(A-A)-l B be a minimal realization of the symbol of equation (1.1), let x be its state space, and let AX = A-BC.
THEOREM 1.1.
In order that for each f E Lm[O,OO) equation (1.1) has a unique solution p ~ E Lm[D,oo) it is necessary and sufficient that p
(i)
AX has no real eigenvalues,
(ii)
x
x
= M $ M ,
where M (resp., MX) is the spectral subspace corresponding to the .
x
eigenvalues of A (resp., A ) in the upper (resp., lower) half plane. In that case for a given f E Lm[O,OO), the solution of (1.1) is given by 00
~(t)
= f(t)
where (1.2)
+
f
p
y(t,s)f(s)ds,
o :,;
t <
00,
o
, _itAx iSA x 1Ce TIe B, 0:,; s < t < 00, { x x _iCe-itA (I_TI)e isA B, o :,; t < s <
yet,s)
00,
x
with TI being the projection of x along M onto M • Let us illustrate the above theorem for the following simple equation: (1.3)
~(t)
-
~
00
f
e-lt-sl~{S)dS
= f (t) ,
o :,;
t <
00.
0
Here c is a positive real constant and the given function f, the unknown
58
M.A.KAASHOEK
~
and the kernel k are scalar functions. To solve (1.3) is to construct a realization of its symbol:
the first step
Put
[~
A
Then wO.)
C
[-c
oJ.
1 + C(A-A)-l B is a minimal realization and
A
x
A -
BC
x
It follows that A has no real eigenvalues if and only if 0 < c < 1. Therefore, assume 0 < c < 1. Let M and MX be as in Theorem 1.1. One computes that
It follows that ~2 = M $ MX, and thus equation (1.3) is uniquely solvable in L [0,00). The projection IT of ~2 along M onto MX is given by p
IT
1
a+l
1
[
-ia
i] a'
where a = 11-c. Using this in (1.2) one finds that for 0 < c < 1 the resolvent kernel of (1.3) is equal to yet,s) =
2~
{(1_a2)e-a/t-sl - (1_a)2 e -a(t+s)},
where, as before, a = 11-c . Condition (i) in Theorem 1.1 is equivalent to the condition that the symbol weAl admits a (possibly non-canonical) Wiener-Hopf factorization, which, in turn, is equivalent to the requirement that the integral operator I-K corresponding to the left hand side of (1.1) is a Fredholm operator on Lm[O,OO). Furthermore, in that case, using the p
formulas for the factorization indices given at the end of Section 1.7 one deduces: x x dim --x- , dim M n M , codim Im(I-K) dim Ker (I-K) M+M and thus
59
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
index (I-K) where P and P
x
x
rank P - rank P ,
are the Riesz projections corresponding to the eigenx
values of A and A 11.2.
=
in the upper half plane (see formula (1.7.4».
A half space problem in transport theory
Wiener-Hopf integral equations appear in a natural way in various half space transport problems. In fact, the simple equation in the example of the previous section is used to solve the problem ~f radiation diffusion in a one dimensional medium (see [36J, p. 310). The Milne problem, which is one of the classical transport problems in Astrophysics, was the starting point for the original Wiener-Hopf paper [39J. However, in general, the symbols of the Wiener-Hopf equations that appear in transport problems are non-rational. Nevertheless, with some modification, the general approach of the previous section also works in this context. To illustrate this consider the following plane symmetric version of the one-dimensional linear transport equation:
a~
]lax (x,]l) + ~(x,]l) =
(2.1)
f1
g(]l,]lI)~(X,]lI)diJ',
-1 ~]l ~ 1, x ~ O.
-1
One thinks of (2.1) as a balance equation which describes the timeindependent transport of particles through a homogeneous, semi-infinite medium. The unknown function ~ represents an averaged density of the particles. The integral in the right hand side of (2.1) describes .the scattering in the medium and the corresponding scattering kernel g is real symmetric. The variable x is the optical depth, which is a position coordinate inside the medium. Further, ]l = cos ~, where ~ describes the direction of the velocity vector. The problem is to find the unknown function ~ under certain boundary conditions. Here we shall consider the so-called half space problem in which the unknown function ~ is required to satisfy (2.2)
with
~+
given, and a certain growth condition for x
+
00
which we shall
specify later. Condition (2.2) tells us that the density of the incoming particles is known. By writing ~(x) (]l) = ~(x,]l), we may consider the unknown function ~ as a vector function on 0 ~ x ~ 00 with values in the Hilbert space L2 [-1,lJ. This allows one to rewrite equation (2.1) as an operator differential equation: (2.3)
T
~(X) + ~(x)
=
BO~(X) ,
o :;
x <
00,
Where the derivative is taken with respect to the norm in L2 [-1,lJ. The
60
M.A.KAASHOEK
operators T and BO' which act on the space L 2 [-1,lJ, are defined by 1
(Tf) (].l)
].If(].l),
(BOf)
(].l)
f
=
g(].l,].l')f(].l)d].l.
-1
Note that because of the symmetry conditions on the scattering kernel g the operator BO is selfadjoint. In what follows we shall assume that I-B O is strictly positive, which in the transport theory means that we are dealing with an absorbing medium. The equation (2.3) has to be considered together with the boundary (2.4)
Here P+ is the orthogonal projection of L2 [-1,lJ onto the space consisting of all functions that vanish a.e. on [-l,OJ and ~+ is a given function in L2 [-1,lJ which is zero on [-l,OJ. If one assumes, as a boundary condition at infinity, that the unknown function Wis (Bochner-) integrable on [0,00), then (2.3) together with the boundary condition (2.4) is equivalent to a WienerHopf integral equation: 00 (2.5)
w(x) -
f
o
k(x-y)w(y)dy
with an operator-valued (2.6)
W(A)
=I
L 1 -kernel,
- (I-iAT)
-1
o
= f(x),
$
x < 00,
of which the symbol is equal to ~
BO'
< A < 00.
The right hand side of equation (2.5) is given by
f (x) (].l)
{ ~ + (].l) exp( - fr) , o
,
-1 $ ].l < O.
To solve (2.5) we want to construct, if possible, a right canonical factorization of the symbol W(A). To do this we employ the method of Theorem 1.7.1 although here W(A) is not rational. First of all we rewrite the symbol W(A) as a transfer function by setting: 'V
W(A) = W(-
1 x-)
(I-B O)
-1
= I + C (A-A)
-1
B,
where (2.7)
A =
-iT,
C
=
A.
Note that the invertibility of I - BO is guaranteed by our hypothesis that I - BO is strictly positive. The operators A, Band C are operators
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
61
acting on the infinite dimensional Hilbert space L2 [-1,1] and the associate operator
x A
(2.8)
=A
- Be
=
(I-B O)
-1
A.
x Next, we want to find the spectral subspaces M and M . The main x
difficulty here is that the spectra of A and A both cut through the real line at zero. However, since A = -iT it is natural to take for M the spectral subspace of T corresponding to the spectrum of T on the negative part of the real line. In other words, we take M = Ker P+. Since I - BO is strictly positive, the operator iA
x
is selfadjoint with
respect to the equivalent inner part 1
[h 1 ,h 2 ] =
f
-1
((I-B O)h 1 ) {t)h 2 {t)dt.
x So there is also a natural choice for the space M , namely the spectral x subspace of the selfadjoint operator iA corresponding to the positive x part of the real"line. Let M and M be chosen in this way. Then the next important step is to prove the decomposition: (2.9)
Assume (2.9) holds true. Then W{A) admits a canonical factorization relative to the real line and the factors can be constructed in precisely the same way as in Theorem 1.7.1. It follows that W{A) admits a canonical factorization. In fact, let IT be the projection of L2 [-1,1] x along M onto M and put -1
I - iATIT{I-iAT)
W+{A)
I - BO - iAT{I-iAT)-l{I-IT)B O•
-
Then W{A)
-1
W (A)
= W_{A)W+{A)
(I-B O)
BO'
is a right canonical factorization of the symbol
W{A). On the basis of this factorization, using the classical method described in Section 1.1, it is not difficult to prove that equation (2.5) has a unique (Bochner-) integrable solution ~ on 0 $ x < 00, which is given by -1 -xT (I-B O) ~(x) e IT+. The matching of the subspaces in (2.9), which is the main problem here, can be proved for a large class of scattering kernels g. For example, if n g
(ll, II ' )
~
j=O
a j P j (ll)Pj (ll') ,
62
M. A.
where p.
J
(~)
KAASHOEK
is the j-th normalized Legendre polynomial and a. < 1 J
(j = O,l, ••. ,n), then (2.9) holds true (see, e.g., [26J). In [33J the decomposition (2.9) is proved for all scattering kernels g such that I - BO is strictly positive and BO
=
ITlaD for some 0 < a < 1 and some
bounded operator D on L2 [-1,lJ. 11.3.
Wiener-Hopf equations with analytic symbols
In this section the symbol weAl of the equation (1.1) is assumed to be analytic in a strip around the real line, i.e., for some positive number h the function weAl is analytic on 11m AI < h. The Wiener-Hopf equation considered in the previous section is an example of an equation with such a symbol. In fact, the symbol weAl in (2.6) is analytic on 11m A I < 1. To extend the theory of rational symbols discussed in Section 11.1 to analytic symbols, the first problem is to realize weAl as a transfer function, i.e., to construct a representation (3.1)
weAl = I + C(A-A)
-1
B,
Since weAl is non-rational one cannot expect a realization (3.1) with a finite dimensional state space. To solve the realization problem one has to distinguish between two cases. First, assume that the symbol weAl is analytic on the real line and at infinity. So there exists an open set ~ on the Riemann sphere such that. the real line and the point infinity are in ~. Let X be the linear space of all ~-valued analytic functions on ~ such that II fez) II Illflll
sup ZE~
< co, max(1,llw(z)ll)
and consider the following operators: A: X + X, (Af) (z) = zf(z) - (z-a)f(co); B: a:m + X, (By) (z) = (z-a)[W(z)y - yJ; C: X + a:m, Cf = f(co). Here a is an a~bitrary complex number in a: \ ~. Then the space X endowed with the norm 111.111 is a Banach space, the operators A, Band Care bounded linear operators, ~ \ {co} belongs to the resolvent set peA) of A and (3.2)
weAl = I + C(A-A)
-1
B,
A€
~ \
{co}.
In particular, the operator A in (3.2) has no spectrum on the real line. We conclude that a symbol weAl which is analytic on the real line and at infinity admits a representation of the form (3.1) with A, Band
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
63
e bounded linear operators, where the operator A, which act~ on an infinite dimensional Banach space X, has no spectrum on the real line. As soon as one has such a representation for the symbol one can analyse equation (1.1) in essentially the same way as in the rational case. For M one takes 1m P, where P is the Riesz projection corresponding to the part of the spectrum of A in the upper half plane, and one proves that equation (1.1) is uniquely solvable in Lm[O,OO) if and only if AX = A-Be p x x has no spectrum on the real line and X = M eM, where M = Ker P and x x P is the Riesz projection of A corresponding to the part of the spectrum in the upper half plane. The formula for the resolvent kernel is then the same as in the rational case. Next, we drop the condition of analyticity at infinity. Then, in general, the symbol weAl cannot admit a representation (3.1) with A, B and e bounded. Indeed, if (3.1) holds true with A, Band e bounded linear operators, then the Neumann expansion of the resolvent of a bounded linear operator implies that weAl is analytic at infinity. So, in this case the best one can hope for is a realization with unbounded operators. For the symbol of the Wiener-Hopf equation in the transport problem of Section 11.2, a realization with unbounded operators comes automatically from the data of the problem. In fact, (2.6) may be rewritten in the form: WeAl = I - iT
-1
(A+iT
-1 -1
)
BO'
where T- 1 is the unbounded inverse of T. In the general case the realization has to be constructed. Here we shall present the solution of the realization problem for a symbol that is analytic in a strip around the real line under some additional conditions on the kernel k(t). The conditions we shall work with are sufficient but not necessary for all considerations that follow. We shall require that on E \ {a} the kernel k has a continuous second derivative which is exponentially decaying and has at most a jump discontinuity at zero. These conditions imply that there exists W > 0 such that (3.3)
j
0,1,2.
In particular, the s¥IDh0l weAl of the corresponding Wiener-Hopf equation is analytic in 11m AI < w. Now, let X be the linear space of all ~-valued functions f on
E \ {a} such that eWitif(t) is bounded and uniformly continuous on E \ {a}. The space X endowed with the norm
III f III
sup e wt IIf(t) II t#O
is a complex Banach space. Let A be the unbounded operator 1d~ on X.
64
M. A. KAASHOEK
Thus the domain V(A) of A consists of all f € X such that f is differentiable and f' € X, and for such a function Af = if'. Further, define B:
,fI1
+ X,
(By) (t) = k' (t)y, 00
C: X + a;m,
Cf = i f f (t) dt. _00
The operator A is a densely defined closed linear operator such that the strip 11m AI < W is in the resolvent set of A, the operators Band Care bounded and by a direct computation one checks that (3.4)
I + C(A-A)-1 B ,
W(A)
11m
AI < w.
Note that the operators CA and AB are also bounded. The main operator A appearing in (3.4) has several interesting properties. To describe these let P be the projection of X defined by (Pf) (t)
=
f (t)
(t < 0),
(Pf) (t) = 0
(t
> 0).
The projection P commutes with A, and hence relative to the decomposition X = 1m P $ Ker P the operator A can be written as an 2x2 diagonal operator: (3.5)
A
0 1 (1m P A+
[A_
o
$
It turns out that -iA+ and iA
Ker P + 1m P
$
Ker P).
are generators of exponentially decaying
Co-semigroups. In fact: (e -i~f) (s)
f (t+s)
(t,s > 0),
(e -itA_f) (s)
f (t+s)
(t,s < 0).
Any closed densely defined operator A on X of the form (3.5) with -iA and iA generators of exponentially decaying Co-semigroups will + be called a bigenerator. If A is a bigenerator, then the projection P of X that induces the partioning (3.5) is uniquely determined by A and is called the separating projection. A bounded linear operator A on X is a bigenerator if and only if A has no spectrum on the real line, and in that case the correspon~ing separating projection is precisely the spectral projection associated with the part of the spectrum of A in the upper half plane, that is, P = __1__ . 2m.
f
r
(A-A)-1 dA ,
where r is a suitable contour in the upper half plane around the spectrum of A in the upper half plane.
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
65
In order that the symbol wO.) admits a realization
THEOREM 3.1. (3.6)
-CO
< A<
00,
where A is a bigenerator and B and c are bounded linear operators such that AB and CA are bounded it is necessary and sUfficient that on ~ \ {a} the kernel k has a continuous second derivative which is exponentially decaying and has at most a jump discontinuity at zero. In that case =
{
k (t)
(3.7)
iCe
-itA . (I-P)B,
_ice-l.tApB,
where
P
t
> 0,
t
< 0,
is the separating projection of
In (3.7) the expression e
-itA
A.
(I-P) has to be understood as the semiqroup
(I-p)e-i~(I-P). Similarly, e-itAp
= pe-itA-p. Theorem 3.1 solves the realization problem. The next step is to use the realization (3.6) to analyse the equation (1.1).
THEOREM 3.2. Let W(A) = I + C(A-A)-l B be a realization of the symbol of equation (1.1), where A is a bigenerator on X and B and C are bounded linear operators such that AB and CA are bounded. Put AX = A - BC. In order that for each f E Lm[O,OO) the equation (1.1) has a unique solution p , E Lm[O,OO) it is necessary and sUfficient that p X
is a bigenerator,
(i)
A
(11)
x =M
X
eM,
with M = 1m P and MX = Ker pX, where P (resp., pX) is the separating projection associated with the bigenerator A (resp., AX). In that case for a given f E Lm[O,OO) the solution of (1.1) is equal to p 00
,(t) = f(t) +
where
J y(t,s)f(s)ds, o
o
$
t
<
00,
X
ice-i(t-S)A (I_px)B + h(t,s), yet,s) • {
with h(t,s) X
along
M
X
_ice-i(t-S)A pXB + h(t,s),
= _iCe-itAx(I_Px ) (I_n)eiSAxpxB x
onto
o
$
s < t <
00,
o
$
t <
00,
S <
and n is the projection of
M •
x In the previous theorem the main difficulty is to prove that A is a biqenerator if (1.1) is uniquely solvable and to construct the
66
M. A. KAASHOEK
corresponding separating projection pX. It turns out that the property "A is a bigenerator" is stable under perturbation of a certain type. THEOREM 3.3.
Let
A
be a bigenerator on x, and let
operator on X such that
~
the strip let
11m AI
(3.8)
sUI? IAI2 11m AI~h
h
V«Ax)2) = V(A 2 ).
be a closed linear
AX
Assume that for some
is contained in the resolvent sets of II (A_Ax )-l -
(A-Al- 1 11
<
A
h > 0
and
AX,
and
00
Then AX is a bigenerator and the corresponding separating projection is given by X
P x
Now let W(A) be as in Theorem 3.2, and assume that equation (1.1) is uniquely solvable in Lm[O,OO). Then there exists a strip around the . p real line such that for A in this strip W(A) is invertible and (3.9)
X -1
(A-A)
-
(A-A)
-1
(A-A)
-1
-1
BW(A)
C(A-A)
-1
•
By using the boundedness of AB and CA, one can deduce from (3.9) that X
the evaluation (3.8) holds true, which yields that A
is a bigenerator.
11.4. Wiener-Hopf equations and time-invariant linear systems So far the results discussed were based essentially on factorization. In the present section we come to a method in which factorization does not appear at all. As before we deal with the Wiener-Hopf integral equation (1.1). In Section 1.1.1 the Fourier transform has been used to reduce equation (1.1) to the vector-function equation: (4.1)
...00
<
A<
00.
Here W(A) is the symbol of equation (1.1). The functions F+(A) and ~+(A) are the Fourier transforms of the given function f and the unknown ~, respectively, where both f and ~ are assumed to be zero on the negative part of the real line. The function F (A) is the Fourier transform of a function fO with support on t ~ O. Now, let us assume that W(A) is rational. Then W(A) admits a realization: (4.2)
W(A)
I + C(A-A)
-1
B,
...00
<
A<
00,
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
67
with a finite dimensional state space ~n and A has no real eigenvalues. Introduce the auxiliarly function
Then (4.1) can be rewritten in the form \X(A) = AX(A) - B~ (A), {
= -CX(A)
F+(A)
+ :+(A) - F (A).
By applying the inverse Fourier transform one obtains the linear system: x(t)
-iAx(t) + iB¢(t),
f(t)
-Cx(t) + ¢(t) - fO(t),
{
which has to be considered on the real line. Recall that we are interested only in functions on the positive half line. By restricting to t ~ 0 one can drop the unknown function fO which is zero on t ~ O. So we are led to the system {
(4.3)
x(t)
-iAx(t) + iB¢ (t),
t
~
0,
f(t)
-Cx(t) + ¢(t),
t
~
0,
A further analysis shows that (4.3) has to be considered with the following. boundary condition at zero: (4.4)
(I-P)x(O) = 0,
where P is the spectral projection of A corresponding to the eigenvalues in the upper half plane. The system (4.3) together with the boundary condition (4.4) is equivalent to equation (1.1) in the following sense: If for a given f
E
Lm[O,OO) the function ¢ p
(1.1), then there exists a unique x
E
E
Lm[O,OO) is a solution of p
Ln[O,OO) satisfying (4.4) such that p
(4.3) with input ¢ has output f, and, conversely, if there exists :XE
Ln[O,OO) satisfying (4.4) such that (4.3) with input ¢ E Lm[O,OO) has p
p
output f, then ¢ is a solution of (1.1). By using this equivalence the first step to solve equation (1.1) is to interchange in (4.3) the role of the input and the output. This leads to the following system: (4.5)
{
Ht)
-i(A-BC)x(t) + iBf(t),
t
~
0,
¢(t)
Cx(t) + f (t),
t
~
0,
which has to be considered together with the boundary condition (4.4) •
68
M.A.KAASHOEK
Next, one uses the variation of constants formula to find the solutions of the first equation in (4.5) that are in Ln[O,OO) and satisfy the p
boundary condition (4.4). In this way one obtains conditions for the solvability of (1.1), formulas for its solutions and various other characteristics of equation (1.1) in terms of the operators A, Band C. As an example we mention the following theorem.
Let weAl I + C(A-A)-l B be a reaLization of the symboL of equation (1.1) with state space ~n~ and assume that A and AX have no x reaL eigenvaLues. Let p (resp.~ p ) be the spectraL subspace of A x (resp.~ A ) corresponding to the eigenvaLues in the upper haLjpLane. Then~ given f € Lm[O,OO)~ equation (1.1) is soLvabLe in Lm[O,OO) if and onLy if p p THEOREM 4.1.
00
J
o
e
isAx
Bf(s)ds
x
€
1m P + Ker P ,
and in that case the generaL soLution of (1.1) is equaL to .
x
t
I
~(t) = f(t) + ce-~tA x + i
x
.
ce-~(t-S)A Bf(s)ds,
o where x is any v2ctor in 1m P such that x + i
co
I
o
e
iSAX
Bf(s)ds
is in
Ker pX. In particuLar~ if T is the integraL operator on Lm[O,OO) defined by p 00
~(t) -
(T~) (t)
I
o
o
k(t-s)~(s)ds,
~
t
<
00,
then Ker T
= {~ I ~(t) = ce- itA 00
1m T
= {f I I
o
e
i~x
x
Bf(s)ds
x, €
x
1m P n Ker P }, x 1m P + Ker P }, X €
and T is a FredhoLm operator with index x index T = rank P - rank P •
One of the main features of the method described here is that it can also be applied in cases when the factorization method cannot be used. For example, the reduction of a Wiener-Hopf integral equation with rational symbol to a linear system can be repeated for convolution equations on a finite interval, T
(4.6)
~(t) -
f
o
k(t-s)~(s)ds
= f(t),
and leads to a system of the form:
0 s t ~ T,
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
0 5 t 5 T,
{ ,'t) = -iAx,t) + i",t), (4.7)
f(t)
=
-Cx(t) +
=
(I-P)x(O)
0,
~(t),
PX(T)
69
0 5 t 5 T,
=
O.
The system (4.7) and the integral equation (4.6) are equivalent whenever the kernel k(t) of (4.6) admits the following representation: (4.8)
k(t)
{
iCe
-itA
. (I-P)B,
_ice-l.tApB,
t > 0, t < 0,
where A, Band C are matrices of sizes nXn, nXm and mXn, respectively, and P is a projection on ~n which commutes with A. If k(t) can be extended to an integrable function on the real line of which the Fourier transform K(A) is rational, then a realization of the symbol I-K(A) yields a representation (4.8) in which A has no real eigenvalues and P is the Riesz projection corresponding to the eigenvalues of A in the upper half plane. By using the equivalence between (4.6) and (4.7) the invertibility properties of (4.6) can be derived from a study of the inverse system:
{
x(t)
= -i(A-BC)x(t)
~(t)
= Cx(t) - f(t),
(I-P)x(O),
PX(T) =
+ iBf(t),
o o
5 t 5 T, 5 t
5 T,
o.
As an example we mention the following theorem.
Asswne that the ke1'nel k(t) of the equation (4.6) admits the representation (4.8), where P is a projeation aommuting with A. . . PUt Ax = A-BC. Then for eaah f E Lm[ O,T J equat~on (4.6) ha s a un~que p solution ~ E Lm[O,TJ if and only if the operator p
THEOREM 4.2.
1m P
-)0
1m P
1m P
is invertible. In that aase the solution of (4.6) is given by t"
~(t)
f(t) +
I
y(t,s)f(s)ds,
o
where y't,s)
={
x x ice-itA TIe iSA B, s < t, x _itA iSA x -iCe (I-TI)e B, s > t,
and TI is the projeation of ~n onto 1m P defined by
M.A.KAASHOEK
70
TIx
=x
- S
-1
Pe
iTA _iTA X e x,
The interplay between systems and integral operators will be the main topic of the third lecture. 11.5.
Further remarks and references
The main reference for Section 11.1 is §4.5 in [3]. The final formula for the resolvent is taken from [4]. Transport theory is a branch of mathematical physics which originated from and has many points of contact with astrophysics and nuclear physics (see, e.g., the recent book [30] and the references given there). For Section 11.2 the main source is [3], Chapter 6, but see also [33], which has, among many other things, the final formulas of the solution and of the resolvent kernel. The importance of the matching of the subspaces in (2.9) was known in transport theory and was proved, for example, in [26] (see also [25]) in connection with another method to solve half space problems. The solution of the realization problem for symbols that are analytic on the real line and at infinity, which is taken from [4], is closely related to the idea of linearization (see [20]); in fact the formulas for A, Band C in (3.2) come from the linearization theorem in [11] (see also [3], Sections 2.3 and 2.4). Symbols that are not analytic at infinity form the main topic of [9]; Theorems 3.1 - 3.3 are taken from [9]. The connections between convolution equations and time invariant linear systems, discussed in Section 11.4, have been developed in [4, 5]. The discrete versions of the results of Sections 11.1 and 11.4 can be found in [3, 4, 5].
III.
SYSTEMS WITH BOUNDARY CONDITIONS AND INTEGRAL OPERATORS
In this third lecture integral operators are viewed as transfer operators of time-varying linear systems with boundary conditions. For such systems the concepts of realization, similarity, and cascade connection will be discussed. Invertibility of integral operators is studied in terms of inversion of systems and factorization in terms of decoupling. On a finite interval the integral operators involved are compact and have a semi-separable kernel. The half line case concerns a large class of non-compact integral operators which includes WienerHopf integral operators with rational symbols. 111.1.
Integral operators as transfer operators
On a finite interval a time varying linear system with boundary conditions has the following state space representation:
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
X(t) = A(t)x(t) + B(t)u(t), { e yet) = e(t)x(t) + D(t)u(t), N1x(a) + N2 x(b)
71
a
:s;
t
:s;
b,
a
:s;
t
:s;
b,
O.
Here A(t) is an nXn matrix function of which the entries are integrable on a :s; t :s; b. The input coefficient B(t) and the output coefficient e(t) are matrices of sizes nXm and mxn, respectively, and the entries of B(t) and e(t) are square integrable on a :s; t :s; b. (For simplicity, the number of input variables is taken to be equal to the number of output variables.) The extePnaZ coefficient ott) is assumed to be measurable and essentially bounded on a :s; t :s; b. We shall refer to A(t) as the main coefficient. The boundary value matrices are constant and have size nXn. In what follows we shall represent the system e by its four coefficients and two boundary value matrices and write e = (A(t) ,B(t) ,e(t) ,O(t) iNl ,N 2 ). To ensure that in e the state x(t) is uniquely determined by the input u(t), the boundary conditions of e are assumed to be weZZ-posed which means that the homogeneous equation itt)
= A(t)x(t),
a
:s;
t
:s;
b, N1x(a) + N2X(b)
= 0,
has no non-zero solution. The well-posedness of the boundary conditions implies that the output y is a linear function of the input u, i.e., the input/output behaviour of the system is described by a linear operator Te' y = Teu, which we sha~l call the transfer operator of e. To determine the transfer operator of e one uses the fundamentaZ matrix of the system e, which, by definition, is the unique absolutely continuous solution of the matrix differential equation: (1.1)
U(t) = A(t)U(t), a
:s;
t
:s;
b, Uta) = I,
where I is the nXn identity matrix. In terms of the fundamental matrix well-posedness is equivalent to 0.2)
By using the variation of constants formula one determines the state x as a function of the input u. By inserting the formula for the state in the second equation of the system one computes that the transfer operator is an integral operator b a :s; t :s; b, Y (t) = (TeU) (t) = O(t)u(t) + f k(t,s)u(s)ds, a with kernel (1. 3)
k (t,s)
{"(t'U(t' ('-P'U(5,-'.(5', -C(t)U(t)PU(s)-l B(s),
aSs < t S b, a S t < s S b,
72
M. A. KAASHOEK
The operator P, which plays an important role in what follows, is well-defined because of (1.2). Using P the boundary conditions of 8 can be rewritten in the form: (I-P)x(a) + PU(b)-l x (b)
0,
and for that reason P is called the canonical boundary value of 8. If 8 is a causal system (i.e., Nl = I, N2 = 0), then P for an
anticausal system (i.e., Nl
= I)
0, N2
operator
the operator P
0, and
=
I. In
general, P is not a projection; in fact (1.4)
From the integrability conditions on the input and output coefficients it follows that the kernel (1.3) of the transfer operator is a Hilbert-Schmidt kernel, and hence we shall consider T8 as an operator on L~[a,b]. The time invariant case (when the coefficients do not depend on t) is of special interest. For example, consider ( x(t) = Ax(t) + Bu(t), (1.5)
,j
l
yet) = Cx(t) + Du(t), N1 X(0) + N2 X(T)
O~t~T,
o
~
t
~ T,
= o.
The fundamental matrix of (1.5) is U(t) e the transfer operator of (1.5) is given by
tA
and hence the kernel of
O~s
(1.6) O~t<S~T,
h were Po
=
( N1+ N 2 e TA) -I N2 e TA
' , h A, If , 1n a dd"1t1on, Po commutes w1t
then kO(t,s) depends on the difference of the arguments only and the corresponding transfer operator is a convolution operator. In fact, if POA = APO' then kO(t,s) = h(t-s), where
{
Ce
h(t) "
tA
(I-PO)B,
-ce~oB,
t > 0, t < O.
Conversely, if the transfer operator of (1.5) is a convolution operator and the triple (A,B,C) is minimal in the sense that
73
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
n-l
n
n-l Ker CA j = (0),
V
1m AjB =
a: n ,
j=O
j=O
then Po commutes with A. If Po is a projection commuting with A, then the boundary conditions of (1.5) can be rewritten as (I-P)x(O) = 0,
PX(T)
=
0,
and we come to systems that were considered in Section 11.4. To see which integral operators on L~[a,b] can be realized as transfer operators, note that the kernel k in (1.2) is semi-separable, i.e., it admits a representation of the form:
(1. 7)
k(t,s) = {
Fl (t)G 1 (s),
a :-:; s < t :-:; b,
-F 2 (t)G 2 (s),
a :-:; t <
S
:-:;
b,
where for V = 1,2 the matrix functions Fv (') and Gv (.) are square integrable on [a,b] and have sizes mxnv and nvxm, respectively. For example, to write (1.2) in the form (1.7) one may take C(t)u(t) (I-P),
u (t) -l B (t)
C(t)U(t)P,
U(t) -l B(t).
Conversely, if T on
,
m
L2 [a,b] is an integral operator, b
(T<j» (t) = O(t)<j>(t) +
f
k(t,s)<j>(s)ds,
a :-:; t :-:; b,
a
such that k is semi-separable and OCt) is measurable and essentially bounded, then T is the transfer operator of a system e, in which case e is called a realization of T. In fact, if k is given by (1.7), then T = Te with (1.8)
e
(A(t) ,B(t) ,C(t) ,O(t); I-Q, QU(b) -1),
where A(t) is an arbitrary integrable nXn matrix function, U(t) is determined by A(t) via (1.1) and B(t)
U(t)[G 1 (t)j, G2 (t)
C(t) = [F 1 (t)
Q = [00
F 2 (t)] U(t)-l
Note that the canonical boundary value operator P of (1.8) is equal to Q, and hence P is a projection. In (1.8) we may take A(t) 0, which
=
M.A.KAASHOEK
74
=
implies that U(t) I. So the transfer operator always has a realization 8 of which the main coefficient A(t) is identically zero and the canonical boundary value operator is a projection. The different choices of A(t) in (1.8) correspond to system similarities. Two time varying systems 8 1 and 8 2 ,
v
1,2,
=
are called similar (notation: 8 1 ~ 8 2 ) if D1 (t) = D2 (t) a.e. on a ~ t ~ b and there exist an invertible nXn matrix E and an absolutely continuous nXn matrix function S(t), a ~ t ~ b, such that det S(t) F 0, a ~ t ~ b, and S(t)A 1 (t)S(t)-1 + S(t)S(t)-1, S(t)B 1 (t),
C1 (t)S(t)-1,
EN;ll S (a) -1,
EN~1)S(b)-1,
almost everywhere on a ~ t ~ b. This notion of similarity appears in a natural way when in 8 the state x(t) is replaced by z(t) = S(t)x(t). Well-posedness of the boundary conditions is preserved under similarity and similar systems with well-posed boundary conditions have similar canonical boundary value operators and the same transfer operator. 111.2.
An application: inversion
Consider the integral equation: b
(2.1)
J
k(t,s)
f(t),
=
a
~
t
~
b.
a
We assume that in (2.1) the kernel is semi-separable; so, in other words, the kernel k is the kernel of the transfer operator of a system 8 with well-posed boundary conditions: x(t) { 8 f(t)
=
A(t)x(t) + B(t)
a
=
C(t)x(t) +
a :-; t
~
t :-; b, ~
b,
o.
N1x(a) + N2x(b)
The problem is to solve (2.1) in L~[a,b]. To do this, we use the equivalence between 8 and (2.1). First we interchange in 8 the role of the input and the output, which leads to the inverse system: x(t)
(A(t) - B(t)C(t»x(t) + B(t)f(t),
a
-C(t)x(t) + f(t),
a :-; t :-; b,
N1x(a) + N2x(b)
=
o.
~
t :-; b,
75
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
Note that the coefficients of e
x
satisfy the desired integrability
conditions. It turns out that the boundary conditions of eX are wellposed if and only if Te is invertible, and in that case (Te)
-1
= Tex.
X
X X Let u (t) denote the fundamental matrix of e ; we shall refer to U (t) as the assoeiate fundamental matrix of e.
Let k be the kernel of the tr>ansfer> oper>ator> of the system e = (A(t) ,B(t) ,C(t) ,I;N1,N2~ and let UX(t) be the fundamental matr>ix of e. Then, given f E L~[a,b~ equation (2.1) has a unique solution ~ E
THEOREM 2.1.
L~[a,bJ if and only if
~ 0, and in that ease
det(N 1 + N2 Ux (b» b
~(t)
= f(t)
J y(t,s)f(s)ds,
-
a ~ t ~ b,
a
with r>esolvent ker>nel =
{
y(t,s)
XX x -1 C(t)U (t) (I-P )U (s) B(s),
s < t,
x x x -1 -C(t)U (t)P u (s) B(s),
s > t,
Note that in the above theorem the invertibility of an integral operator acting on an infinite dimensional linear space is reduced to the x invertibility of a finite matrix. The matrix N1 + N2 U (b) can be used to derive other invertibility properties of equation (2.1). For example, for f E L~[a,bJ equation (2.1) is solvable in L~[a,bJ if and only if x b x -1 x N2 U (b) U (s) B(s)f(s)ds E Im(N 1 + N2 U (b». a x The connections between N1 + N2 U (b) and equation (2.1) are quite
J
deep. To illustrate this further, let us put a spectral parameter (2.1) :
~
in
b
(2.2)
~(t) + ~
f
k(t,s)~(s)ds = f(t),
a
x
a
~
t
~
b.
x
To deal with (2.2) one has to replace U (t) by U (t;~) which, by definition, is the unique absolutely continuous solution of
d~ UX(t;~)
=
(A(t) -
~B(t)C(t»Ux(t;~),
{ U
Put
S(~)
(a;~)
= I.
x N1 + N2 U (b;~). The function
S(~)
is an nXn matrix function
76
M. A. KAASHOEK
of which the entries are analytic on the full complex plane. One can prove that the integral operator I + ~K defined by the left hand side of (2.2) and the analytic matrix function S(~) are related in the following way:
(2.3)
[ I+~K o
1=
0
E(~)
s (~) [ 0
~
E
a:.
Ia:n
Here E(~) and F(~) are invertible operators which depend analytically on the parameter ~ and which can be described explicitly in terms of the coefficients A(t), B(t), C(t) of the realization 8. The analytic equivalence (2.3) allows one to derive the spectral data of the integral operator K (i.e., its eigenvalues, eigenvectors, generalized eigenvectors, Jordan chains, etc.) from the corresponding data of the matrix function S(~). For this reason S(~) is called the ind~cator of the integral operator K. 111.3.
Integral equations with a semi-separable kernel on the half line
In this section we consider the case of a semi-infinite time interval. We start with an integral equation 00
(3.1 )
~(t) +
f
k(t,s)~(s)ds = f(t),
O!>t
o
of which the mXm matrix kernel admits the following semi-separable representation:
(3.2)
k(t,s) = {
c (t) u (t) (I-P) u (s) -1 B (s)
,
-1
-c (t) u (t) PU (s)
B (s) ,
o
!> s < t <
00,
o
!> t < S <
00
Here U(t) is determined by the differential equation: (3.3)
U(t)
= A(t)U(t),
0 !> t <
U(O)
00,
=
I,
and for simplicity we shall assume that in (3.2) the operator P, which acts on
a: n ,
is a projection. with the representation (3.2) of the kernel
x
we associate an nXn matrix function U (t) defined by: (3.4)
·X
u
(t)
(A(t) - B(t)C(t»Ux(t), 0 !> t <
00,
Ux(O)
1.
x
We shall refer to U (t) as the associate fundamentaZ matrix. Assume that in (3.3) the entries of A(t) are integrable on o !> t < 00 and that the entries of B(t) and C(t) are in L 2 [0,OOJ. Then equation (3.1) can be analysed in precisely the same way as equation (2.1) in the previous section. In fact, under the stated integrability
x
conditions, the functions U(t) and U (t) are well-defined and the limits
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
lim t-+oo
u(oo)
x
u (00)
U (t),
lim
U
x
77
(t)
t-+<x>
exist and are invertible matrices, which allows one to repeat the arguments of the previous section. one only has to replace a by 0, b by 00, Nl by I-P and N2 by PU(OO)-l, and one sees that the invertibility properties of (3.1) are determined by the matrix I-P + PU(OO)-l UX (oo). Note that under the above integrability conditions the integral operator K corresponding to the integral in the left hand side of (3.1) has a Hilbert-Schmidt kernel and hence K cannot be a Wiener-Hopf integral operator. To get a class of integral equations of the form (3.1) with non-compact kernels including, for example, some Wiener-Hopf equations, it is necessary to look for other conditions on A(t), B(t) and C(t). In what follows we shall assume that B(t) and C(t) are essentially bounded, that A(t) is locally integrable and that the projection I-P is an exponential dichotomy for the differential equation x(t) = A(t)x(t) , o ~ t < 00. The latter condition means th~t there exist positive constants L, M, a and S such that (3.Sa) (3.Sb)
II II
u (t) (I-P) u (s) -111 u (t) PU (s) -111
~ Le -a (t-s) ,
~ Me -S (s-t) ,
o
~
s < t < 00,
o
~
t < S < 00
Since A(t) is locally integrable, the function U(t) is well-defined by (3.3). If (3.Sa) and (3.Sb) hold true we shall also say that I-P is. an exponential diohotomy for u(t). The inequalities (3.Sa) and (3.Sb) do not determine P uniquely, but only its kernel. In fact, if I-P is an exponential dichotomy for U(t), then Ker P
{x E ~n
=
I
U(t)x E L7[0,00)},
and any projection IT of ~n such that Ker IT = Ker P is again an exponential dichotomy for U(t). Note that our conditions allow that A(t), B(t) and C(t) are -itA independent of t. If A(t) = -iA does not depend on t, then U(t) = e -itA The function U(t) = e has an exponential dichotomy if and only if A has no eigenvalues on the real line, and in that case I-P is an exponential dichotomy for U(t) = e- itA if and only if Ker P is the space spanned by the eigenvectors and generalized eigenvectors of A corresponding to eigenvalues in the lower half plane. It follows that Our class of kernels includes convolution kernels of the form k(t,s) = h(t-s), where h (t)
{
ice-itA(I_P)B,
t > 0,
-iCe -itA --1>B,
t < 0,
M.A.KAASHOEK
78
such that A has no real eigenvalues and P is the spectral projection corresponding to the eigenvalues of A in the upper half plane. The integral equations corresponding to the kernels k(t,s) = h(t-s) are precisely the Wiener-Hopf integral equations with rational symbols. To study the integral equation (3.1) the following equivalence with time varying systems is most useful. THEOREM 3.1. Let the kePnel k(t,s) be of the form (3.2), where B(t) and C(t) are essentially bounded and l-P is an exponential dichotomy
for u(t)
with
U(t)
as
in
(3.3). Then, given f
€
Lm[O,OO), the p
function ~
€
Lm[O,oo) is a solution of (3.1) if
a unique x
€
Ln[O,oo) such that the time varying system
~
p
and only if there exists
p
os:t
itt) = A(t)x(t) + B(t)u(t), { yet) = C(t)x(t) + u(t), (l-P)x(O)
with input u =
~
os:t
= 0,
responses with output y = f.
From Theorem 3.1 and using the idea of cascade connection, it can be proved that the class of all integral operators T on Lm[O,OO) , p
00
(T~) (t) = a~(t) +
f
o
k(t,s)~(s)ds,
os:t
where a € ~ and k(t,s) is any kernel of the type considered in the previous theorem, forms an algebra. To analyse equation (3.1) one interchanges in ~ the role of the input and the output which leads to x
x
the study of the associate fundamental matrix U (t). Note that U (t), which is defined by (3.4), exists, because under our conditions on A(t) , B(t) and C(t) the matrix function A(t) - B(t)C(t) is locally integrable. The following theorems are samples of the results one obtains by analysing (3.1) in terms of ~.
Let the kePnel k(t,s) be of the form (3.2), where B(t) and are essentially bounded and l-P is an exponentially dichotomy for U(t). In order that for any f € L~[O,OO) the equation (3.1) has a unique solution ~ € L~[O,OO) it is necessary and SUfficient that the associate fundamental matrix UX(t) has an exponential dichotomy l_p x such that
THEOREM 3.2. c(t)
~n
=
lm P e Ker pX
In that case the solution
~
is given by
00
~(t) = f(t)
-
f
o
y(t,s)f(s)ds,
o
s: t <
00,
79
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
with resolvent kernel _
XX -1 C(t)U (t) (l-Il)U (s) B(s), x x-1 -C(t)U (t)IIU (s) B(s),
{
y(t,s) -
o
~
s < t
<
00,
o
~
t < s <
00,
where II is the projection of ~n along Ker pX onto 1m P. THEOREM 3.3. Let the kernel k(t,s) be of the form (3.2), where B(t) and C(t) are essentially bounded and I-P is an exponential dichotomy
for U(t). In addition, assume that UX(t) has an exponential dichotomy x m I-P. Then the integral operator l+K on L 1 CO,(0), 00
f
(Kq,) (t) =
o
k(t,s)q,(s)ds,
~
t <
00,
o
is a Fredholm operator, Ker (I+K) = {q,
I
L7CO,(0)
E
x
C(t)U (t)x,
q,(t)
x
X E 1m P n Ker P },
1m (I+K)
{f
00
m
E
If
L 1 CO,(0)
x x
P u (s)
-1
B(s)f(s)ds
o
E X
E
1m P + Ker P },
and its Fredholm index is equal to index (l+K)
rank P
Furthermore, if for f
E
x
rank P.
-
L~CO,oo) the equation (l+K)¢
f
is solvable in
L7CO,00), then one of its solutions is given by 00
=
q,(t)
f(t) -
f
where
x
{
y(t,s)
Here II x
P
x
~
-
x
_C(t)U (t) (l-II)u (s)
-1
~
t < 00,
B(s),
_C(t)U x (t)TIu x (S)-l B (s),
+ (l-P
S = P P: 1m P
o
y(t,s)q,(s)ds,
o
x
)8
+ x
P , with
8
+
o
~
s < t <
o
~
t < s < 00
00,
a generalized inverse of the operator
x
1m P •
Theorem 3.2 and Theorem 3.3 hold true also for L:CO,oo) and with some modifications for LmCO,OO), 1 < P < P
00.
80
M. A. KAASHOEK
III.4.
Cascade connection and factorization of operators
We return to time varying systems with well-posed boundary conditions on a finite interval a ~ t ~ b. For ~ = 1,2 let 8~ = (A~(t) ,B~(t) ,C~(t), O~(t)i
(~)
N1
(~)
,N 2
) be such a system. By definition (cf., Section I.S),
the cascade connection (or product) of 8 1 and 82 is the system 8 (A(t) ,B(t) ,C(t) ,OCt) iN1 ,N 2 ) with
=
B1 (t)C 2 (t)j, A(t)
B(t) A 2 (t)
C(t)
0 1 (t)C 2 (t)],
OCt)
o 1,2.
j
N. J
(2) N. J
We shall write 8
=
8 1 8 2 , It is not difficult to see that 8
=
8 1 8 2 is
the system with well-posed boundary conditions which one gets if the output of 8 2 is equal to the input of 8 1 , It follows that (4.1)
We shall use cascade decompositions to construct factorizations of the operator I+K, when K is an integral operator with a semi-separable kernel. Let 8 be a realization of I+K, i.e., the operator I+K is the transfer operator of 8. Let 8 1 and 8 2 be realizations of the operators I+K 1 and I+K 2 , respectively. Since similar systems have the same transfer operator, it is clear from (4.1) that 8
~
8 1 82 implies that
(4.2) So we are interested in factorizations of 8 modulo similarity. Let 8 = (A(t) ,B(t),C(t) ,IiN 1 ,N 2 ) be a system with well-posed
=
boundary conditions of which the external coefficient OCt) I. Let U(t) be the fundamental matrix of 8, and let P be its canonical boundary value operator. A projection IT of onto
~n
~n1 e ~n2 along ~n1
~n2 is called a supporting projection of 8 if P P
) 12j: ~nl
P 22
e ~n2
+
~nl e ~n2
81
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
and the Riccati matrix differential equation R(t)
=
[1 1
R(t) ]U(t) -l B (t)C(t)U(t)
(-~~t)],
{
(4.3)
R(a) = 0, has a solution R(t): ~n2 + ~nl on a $ t $ b. with the supporting projection IT we associate two systems, namely:
Here R(t) is the solution of (4.3) on a $ t $ b. Let us assume that in what follows all systems have external coefficient equal to I.
Let 8 be a system with well-posed boundary conditions. If IT is a supporting projection of 8, then 8 ~ In(8)r n (8). Conversely, if 8 ~ 8 1 8 2 , then there exists a supporting projection IT such that
THEOREM 4.1.
Let us specify Theorem 4.1 for causal/anticausal factorization. So, consider 8 ~ 8_8+, where 8 is causal and 8+ is anticausal, i.e., the systems 8
and 8+ are of the form
,c 1 (t)
8
(A 1 (t) ,B 1 (t)
8+
(A 2 (t) ,B 2 (t) ,C 2 (t) ,1;0,1 2 ),
,1;1 1 ,0),
Let IT be a supporting projection, and assume that l r n (8)
~
(8) ~ 8 and n 8+. Since similar systems have similar canonical boundary
value operators, one sees that Pll
= 0
and P22
= 12 ,
and hence P12
= 0,
because of the boundary condition at b in the Riccati equation (4.3). It follows that the canonical boundary value operator P is equal to the supporting projection IT. Next, one uses the conditions of solvability of an initial value Ricatti equation to find out what it means that P is a supporting projection. In this way one comes to the following theorem.
Let 8 be a system with well-posed boundary conditions, let U(t) be the fundamental matrix and UX(t) the associate fundamental matrix of 8. Then 8 ~ 8_8+, where 8_ is causal and 8+ is anticausal, if and only if the canonical boundary value operator P of 8 is a projection THEOREM 4.2.
M. A. KAASHOEK
82
and (4.4)
det (I-P+U(t)
In that case 8 8
~
8 8
- +
(0,
=
-1 x
U (t)P)
~
a ::; t ::; b.
0,
with
(I-P+R(t))U(tJ- 1 B(t), C(t)U(t)
I
'
I; I Ker P' 0),
IKer P (0, PU(t)-l B (t), C(t)U(t) (P-R(t))i
8+
I; 0, 11m p), 1m P
where (4.5)
R(t) = (I-P) (I-P+U(t)
-1 x
U (t)P)
-1
a ::; t ::; b.
P,
Let us apply Theorem 4.2 to integral operators. Let K be an integral operator with a semi-separable kernel. Choose (which is always possible) a realization 8 of I+K of which the canonical boundary value operator P is a projection. Let U(t) be the fundamental matrix of 8 and UX(t) its associate fundamental matrix. Assume (4.4) holds true. Then formula (4.1) and Theorem 4.2 imply that (4.6)
I+K
=
(I+K ) (I+K ),
-
+
where
t
(K _ CP) (t)
1 C(t)U(t) (I-P+R(s))U(s)- B(s)<j>(s)ds,
f a
(K+<j» (t) = -
a
f
C(t)U(t) (P-R(t) )U(s)
-1
B(s)<j>(s)ds,
t with R(t) given by (4.5). Formula (4.6) is a so-called LU-factorization of I+K. It can be proved that, conversely, if I+K admits an LUfactorization, then necessarily (4.4) holds true. The class of time invariant systems is not closed under the factorizations considered in Theorem 4.2, that is, if 8 ~ 8 8 and 8
- +
is time invariant, then it does not follow that 8_ and 8+ are also time invariant. Consider, for example, on system:
8
( [ -1
°
1
0] [-1 l ' 1 j'
[1
° :;
1J, 1;
t
::; 1 the time invariant
[°1
0] [0 0'
°
0] ) 1 .
The canonical boundary value operator P of 8, its fundamental matrix x U(t) and its associate fundamental matrix U (t) are given by:
83
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
P
=
[
°0
OJ l ' U(t)
=
It follows that det (I-P+U(t) Hence 8 ~ 8 8 with - +
[e-°
U
but there is no factorization e
U (t)P)
e
~
x
COS
t
[
(t)
-i sin t
-1 x
8
111.5.
t
-t
e
-t
cos t
#
-i sinttJ. cos
° for °
~ t
~ 1.
1; 1,0),
8- e + with e
and e+ time independent.
Remarks and references
probably the first systematic paper on time varying systems with wellposed boundary conditions is [31J, where one also will find a number of examples of such systems. Note, however, that Sturm-Liouville two point boundary value problems can be put into this context, in which case the Green's function becomes the kernel of a transfer operator. In the analysis we have given the variable t does not have to be a time variable, which allows one to include many non-trivial examples, e.g., transport problems in non-homogeneous media. Sections 111.1 and 111.2 are fully covered by [18J, which also treats discrete time systems with well-posed boundary conditions. For time varying systems the problems of minimality and minimal realization, which are not discussed here, have several different and interesting aspects. One aspect, connected with dilation, reduction and similarity, is completely clarified in [19J. The results of Section 111.3 are taken from [21J. The notion of exponentially dichotomy, which plays an essential role in Section 111.3 (and in [21J) is due to J.L. Massera and J.J. Schaffer [32J (see, also [15, 16J). The discrete analogues of the results of Section 111.3, which require the analysis of the notion of dichotomy for difference equations, also appear in [21J. The coupling method from [7J, which allows one to study various classes of integral operators in terms of simpler operators and often just in terms of matrices, plays an important role in the proofs of the theorems in Sections 111.2 and 111.3. Section 111.4 concerns joint work with I. Gohberg which is still in progress and aims at a complete theory of minimal factorization of integral operators, of which LU factorization is a special case. LIST OF REFERENCES 1.
2.
Adams, G.T.: 'A nonlinear characterization of stable invariant subspaces', IntegraL Equations and Opepatop Theopy ~ (1983), pp. 473-487. Bart, H., Gohberg, I., Kaashoek, M.A.: 'Operator polynomials as inverses of characteristic functions', IntegraL Equations and
84
M A.KAASHOEK
OperatDr Theory 1 (1978), pp. 1-18. 3.
4.
5.
6.
7.
8.
Bart, H., Gohberg, I., Kaashoek, M.A.: Minimal factorization of matrix and operator functions. Operator Theory: Advances and Applications, Vol. 1, Birkhauser Verlag, Basel, 1979. Bart, H., Gohberg, I., Kaashoek, M.A.: 'Wiener-Hopf integral equations, Toeplitz matrices and linear systems.' In: Toeplitz Centennial (Ed. I. Gohberg), Operator Theory: Advances and Applications, Vol. 4, Birkhauser Verlag, Basel, 1982; pp. 85-135. Bart, H., Gohberg, I., Kaashoek, M.A.: 'Convolution equations and linear systems', Integral Equations and Operator Theory ~ (1982), pp. 283-340. Bart, H., Gohberg, I., Kaashoek, M.A.: 'Wiener-Hopf factorization of analytic operator functions and realization', Wiskundig Seminarium der Vrije Universiteit, Rapport nr. 231, Amsterdam, 1983. Bart, H., Gohberg, I., Kaashoek, M.A.: 'The coupling method for solving integral equations'. In: Topics in Operator Theory, Systems and Networks, The Rehovot Workshop (Ed. H. Dym, I. Gohberg). Operator Theory: Advances and Applications, Vol. 12, Birkhauser Verlag, Basel, 1984, pp. 39-73. Bart, H., Gohberg, I., Kaashoek, M.A.: 'Wiener-Hopf factorization and realization'. In: Mathematical Theory of Networks and Systems,
Proceedings of the MTNS-83 International Symposium, Beer Sheva, Israel ted. P. Fuhrmann), Lecture Notes in Control and Information 9.
10.
11.
12.
13.
14. 15. 16.
17. 18.
Sciences, nr. 58. Springer Verlag, Berlin, 1984; pp. 42-62. Bart, H., Gohberg, I., Kaashoek, M.A.: 'Spectral analysis of systems of Wiener-Hopf integral equations with symbols that are analytic in a strip around the real line', in preparation. Bart, H., Gohberg, I., Kaashoek, M.A., Van Dooren, P.: 'Factorizations of transfer functions', Siam J. Control Optimization l~ (1980), pp. 675-696. -Boer, H. den: 'Linearization of operator functions on arbitrary open sets', Integral Equations and Operator Theory l (1978), pp. 19-27. Brodskii, M.S.: Triangular and Jordan representations of linear operators. Transl. Math. Monographs 32, Amer. Math. Society, Providence R.I., 1971. Brodskii, V.M., Gohberg, I.C., Krein, M.G.: 'On characteristic functions of an invertible operator', Acta Sci. Math. (Szeged) 32 (1971), pp. 141-164. Cohen, N.: 'On minimal factorizations of rational matrix functions', Integral Equations and Operator Theory § (1983), pp. 647-671. Coppel, W.A.: Dichotomies in stability theory. Lecture Notes in Mathematics, Vol. 629, Springer-Verlag, Berlin, 1978. Daleckii, Ju. L., Krein, M.G.: Stability of solutions of differential equ.ations in Banach space. Amer. Math. Soc. Transl. 43, American Mathematical Society, Providence R.I., 1974. Foia~, C., Salinas, N.: 'Stably invariant subspaces of Hilbert space operators'. Kansas University, preprint, 1980. Gohberg, I., Kaashoek, M.A.: 'Time varying linear systems with boundary conditions and integral operators, I. The transfer operator and its properties', Integral Equations and Operator
MINIMAL FACTORIZATION, LINEAR SYSTEMS AND INTEGRAL OPERATORS
Theory
19.
20.
21.
22.
23.
24. 25.
26.
27. 28. 29. 30.
31. 32. 33.
34. 35. 36.
Z
85
(1984), pp. 325-391. Gohberg, I., Kaashoek, M.A.: 'Time varying linear systems with boundary conditions and integral operators, II. Similarity and reduction', Wiskundig Seminarium der Vrije Universiteit, Rapport nr. 261, Amsterdam, 1984. Gohberg, I., Kaashoek, M.A., Lay, D.C.: 'Equivalence, linearization and decompositions of holomorphic operator functions', J. Funct. Anal. ~§ (1978), pp. 102-144. Gohberg~ I., Kaashoek, M.A., Schagen, F. van: 'Non-compact integral operators with semi-separable kernels and their discrete analogues: inversion and Fredholm properties', Integral Equations and Operator Theory (to appear). Gohberg, I.C., Krein, M.G.: 'Systems of integral equations on a half line with kernels depending on the difference of arguments', Uspehi Mat. Nauk l~ (1958) no. 2 (80), pp. 3-72 (Russian) = Amer. Math. Soc. Transl. (2) 1~ (1960), pp. 217-287. Gohberg, I.C., Krein, M.G.: Tneory and applications of Volterra operators in Hilbert space. Transl. Math. Monographs 24, Amer. Math. Soc., Providence R.I., 1970. Gohberg, I., Lancaster, P., Rodman, L.: Matrix polynomials, Academic Press, New York N.Y., 1982. Hangelbroek, R.J.: 'A functional analytic approach to the linear transport equation'. Ph.D. Thesis, University of Groningen, The Netherlands, 1973 = Transport Theory and Statist. Phys. 5 (1976), pp. 1-85. Hangelbroek, R.J., Lekkerkerker, C.G.: 'Decompositions of a Hilbert space and factorization of a W-A determinant', Siam J. Math. Analysis § (1977), pp. 459-472. Kailath, T.: Linear Systems, Prentice-Hall, Inc., Englewood Cliffs N.J., 1980. Kalman, R.E.: 'Mathematical description of linear dynamical systems', Siam J. Control 1 (2) (1963), pp. 152-192. Kalman, R.E., Falb, P.L., Arbib, M.A.: Topics in mathematical system theory, McGraw-Hill, New York, 1969. Kaper, H.G., Lekkerkerker, C.G., Hejtmanek, J.: Spectral methods in linear transport theory. Operator Theory: Advances and Applications, Vol. 5, Birkhauser Verlag, Basel, 1982. Krener, A.J.: 'Boundary value linear systems', Asterisque Z~/Z~ (1980), 149-165. Massera, J.L., Schaffer, J.J.: Linear differential equations and function spaces, Academic Press, New York, 1966. Mee, C.V.M. van der: Semigroup and factorization methods in transport theory. ph.D. Thesis Vrije Universiteit, Amsterdam, The Netherlands, 1981 = Math. Centre Tracts 146, Mathematisch Centrum, Amsterdam, 1981. Ran, A.C.M.: semidefinite invariant subspaces; stability and applications. Ph.D. Thesis Vrije Universiteit, Amsterdam, 1984. Rosenbrock, H.H.: State-space and multivariable theory, Nelson, London, 1970. Sobolev, V.V.: A treatise on radiative transfer. Van Nostrand, Princeton N.J., 1963.
86
37. 38. 39.
M.A.KAASHOEK
Sz.-Nagy, B., Foia~, C.: Harmonic analysis of operators on Hilbert space. North-Holland Publ. Co., Amsterdam, 1970. Van Dooren, P.: 'Factorization of a rational matrix: the singular case', Integral Equations and Operator Theory (to appear). Wiener, N., Hopf, E.: 'Ueber eine Klasse singularer Integralgleichungen', Sitze Ber. Preuss. Akad. Wiss., Phys.-Math. Klasse 31 (1931), pp. 696-706.
HA-PLITZ OPERATORS: A SURVEY OF SOME RECENT RESULTS
N.K.Nikol'skii Leningrad Branch, Steklov Math Institute Fontanka 27, Leningrad, 191011, USSR
O. Introduction, and formulae linking Hankel and Toeplitz operators 1. Singular numbers 2. Inverse spectral problems 3. Toeplitz operators and exponential bases 4. Spectral multiplicity problem
O.
INTRODUCTION
Originally, Hankel and Toeplitz operators are defined as operators acting on t~ and having matrices with entries depending only on the sum or, respectively, difference of indices: r={ tn+Kl)t. K~O' T = {t n - K} 11.,K~O • Many close relation~ of such operators (matrices) to various problems of algebra, analysiS, differential equations were discovered as early as in the last century. But the spectral theories of Hankel and Toeplitz operators start their deVelopment only in the late 50th. Now they are joined 87
88
N. K. NIKOL'SKII
within the spectral theory of ~nkel and Toeplitz operators. Por most problems of the theory is is more 9; II~ convenient to take the Pourier transfonD. t - + H ,
)( ={xnJ~~o t ~ X= and T
~ :x:,~v D ~ n~o
on the Hardy space
1.l1tt
and to consider
,
2
H=
{
f
r
~
e
L ('If)
:
A
0, tv -< 0 } ; here 1r = { ~ e; {, : I ~ I = 1 J is the unit circle, ~(n) stands for the n-th Fourier coefficient. Moreover, instead of r we consider the operator H from H? to H: = Lt.C]I') e H~ which coincides with r up to an isometric factor and has the same matr1x(rKTK )J but with respect to the f
(11-) =
pair of bases {zl-t. ll-t. ~ 0
and {l Kf"1 JI(~O, Z
being the
identi ty mapping of 'If, X(~) ~ ~, ; e 71' • In other words, H=1r where}f = Z f (~), ie L?-('il') • It is clear that '} is an involution on L({,('!r> inII~ 112terplacing M and N_ • The following two theorems fonD. a background of the theory. There are many sources where the interested reader can find the proofs as well as numerous comments and subsequent developments concerning these theorems, [9], [10J , [17] , [18J , [22J , [24J , [25J • See also [16] for further information. In the statements written below all operators are supposed to be defined only on the set ~A of all polynomials in % • Theorem (Nehari, 1957). A Hankel operator ~ Ii : ~~- W~ is continuous iff there exists a function
HA-PLITZ OPERATORS
and P-
H~
•
89
stands for the orthogonal projection onto Moreover, I ~
=dt~t (~,~oo) tained.
=
II P-YlII BMo
Here Hoo = Loo n ~~ and BMO A = P+ Loo( 11') =I-P_ •
A
and
in! and diet
are at-
is the usual Hardy algebra, with the guotient norm, P+ =
Theorem (Brown, Halmos, 1963). A Toeplitz operator rr ,1" = Ht.-+- ~ ~ is continuous iff there exists a function <.p E L00 (1r> such that l' = 1<jl ,
J.e H"'. Moreover
II T<jlll= I
Outline of the contents. §1 contains the operator valued version of the well-known Adamyan-Arov-Krein theorem on singular numbers of Hankel operators (following S.R.Treil). It is worth mentioning that a great many of applications are forced to deal with the Ha.-plitz operators acting on spaces of vector valued functions, i.e. the Ha-plitz operators with matrix and operator valued symbols. Among such applications we mention only interpolation problems arising from the system theory, prediction theorems for vectorial stationary processes, spectral theory on the Sz.-Bagy-Foia~ functional model, investigations of systems of the Wiener-Hopt equations, and so on, and so on. In particular, in §1 we need an operator valued analogue ot the Behari theorem. In §2 we explain (after S.R.Treil and V.I.Va-
90
N. K. NIKOL'SKII
syunin) a partial solution of the inverse spectral problem for moduli of Hankel operators I Wq>I=(H~ HIjI) (I", In §3 relations of the spectral theory of Toeplitz operators to some geometric characteristics are exposed (angles between model subspaces, free interpolation, geometric properties of systems of exponentials and of reproducing kernels of model spaces). In §4 some very recent results on the problem of spectral multiplicities for Toeplitz operator are discussed. I am indebted to Prof. S.Khrllshchev for valuable language consultations •
••• We conclude this Introduction with a list of formulae linking Hankel and Toepli tz operators with each other as well as with some other operators. Despite their simplicity most of the formulae will be very useful in what follows. The first four are immediate consequences of the definitions of H~ and
'rljl • 11411" ,
(jl E
L00.
for 'I171E: Hoo,
(j>eLoo •
for ~
feH. (4) i
rtp '1' 11 =
H; HljI
(j>eL00,
91
HA-PLITZ OPERATORS
(5)
H~ H~ =
l'KfI"I, y* +
VH~ H141 V*
for u,e Loo ,
\ u\ = 1 a.e.; here T\.I. = Y I 'T\.(. I is the standard polar decomposition of ~1V (V is a partial isometry) and PE stands for the orthogonal projection onto E , Ec. H~. (Hruacev, Peller; see [24J). Proof. H~~ u = I -Tu, Tu =1- YI rtvl~V*
= 1- Vy*+ Y(I - ITu, I~) y* = = I - yy* + YH~ Hu V*"= = PK~~V* -t- YH~ ~ u, y*. • (6) Ql-Isf=f(M Q ) Pe for ;t€H co
,
=
G being an
inner function; here Ps= PKe , where Ke =def ~ ~ e l'IHt\:1 is a model space and Me is a model operator (the compression of the shift onto Ke ), MQ ~ = PG~~, q,e 1(9 . Proof. Pa=9P- G on H~ and ;f(Me)~=P9;f9on Ka ( f ( Me) being defined for polynomials as usual and extended by continuity to all e H00 ) ,from which the desired formula follows. •
s
(7) Q1 JTe g
1 g,
J G"'"
Ii:
on Qt, = H~ (j) Q1' G~ ; here J;f= Proof. For f
6
Hq.
=
KQ~ 1,r,
IH~ ffi (p" -
I KlI
,,~I7!k
)
for every inner functions ;fe L~(,1I').
E9 K9~ ,
e1Jrp~f~ JQ~f= e1JP+QfQ~J9~.f= GfJ~Jelg~9,.5 = 91 ~Q1 f .•
=
92
N. K. NIKOL'SKII
for 19, \jI e
Leo
and 5-, ~ e H~
(A.L.Volberg).
Proof. Multiply by 1."", 146 Z ,and integrate over 'lP the left and the right hand part of the identity. •
1.
SINGULAR NUMBERS
The singular numbers ~)t. , a ~ 0 space operator A are defined by ~n. (A)
=
in! { II A- F' I : f
bounded, rank
Ii' ~
11,
is linear,
j . ~
The following properties of and easy to prove. (a)
of a Hilbert
~1'l1 ( A) , , ~o ( A) =
"
-numbers are classical
A1/
.
(b) t1.m~"\1, (A) = 0 iff A is a compact opera11< tor (i n s y m b 0 1 s : A€ f'00 ). (c) ~"" (A> is the 14-th point from the right of the spectrum of the modulus I AI = (A * A) 1/" (eigenvalues are counted with multiplicities, the first point of the continuous spectrum t>eo ( A) is counted as an eigenvalue of infinite multiplicity). (d) If Ce'r'eo t then C=E ~l'\,(C) (.,X""')1/n. ( { 11.~
HA-PLITZ OPERATORS
93
Theorem (Kronecker. 1881) • f-f
where
at~
is the set of all rational functions with
poles in the unit disc II) and with deg ~ 11, Both inf and dist are attained.
,
a~0 •
The second equality is nothing but the Behari and the Kronecker theorem. By the way, one can interpret this equality as ~1\- ( l-ttp) = the best B M0A -approximation of ~
"tI, (p_ tp ) ) ,
and this formula permits one to apply Hankel operators to the investigation of the relationship between smoothness properties of functions and the rate of decreasing of their best rational approximations {"I.'\1.1 • The most important and compl.ete results in
94
N. K. NIKOL'SKII
this direction were obtained by V.Peller [23J; see also a recent survey [22] for a discussion and more details. Among many other consumers of theorems on :, -numbers of Bankel operators we mention here only a) interpolation problems by rational functions and especially those arising from circuit-theoretic problems (see [1] as an origin and [11J for a survey of new methods and applications), and b) prediction theory for stationary Gaussian processes and in particular the investigation of mixing properties of a process in terms of the spectral measure (see [23]). In the next section we touch briefly the case b). As to a) it should be noted that most of these applications require a vector-valued version of the AAK characterization of singular numbers. Before turning to a vector-valued generalization of AAK theorem recall that its predecessor, the Nehari theorem, admits a generalization of the following form (see [21J, [2]). Below things are going in the vector valued W~-spaces H~(E), H:<E) ( c:: La~ ( E» on the unit circle. By a Hankel operator from ~.,( Ef ) to H~ ( E?" ) we mean an operator on the set of E~-valued polynomials which has a Hankel matrix (fa-rk ) with respect to the bases { ~~Ed 'K-?; 0 ' {1;I(+" E~JI(~o ; [, ' Ef, are some auxiliary Hilbert spaces. Theorem (Page. 1910). A Hankel operator H from U~l E.,) to U~ (E~) is continuous iff there exists a bounded operator-valued E~ ---- E~ function q> on the circle 'll' (1 n s y m b 0 1 s q>eLoo<E.,-E:I,» such that 1-1 = H
HA-PLITZ OPERATORS
95
~qlf dei- P-4> £-, 5e. Ht( E1 ) Moreover,
II H~II
=
cli~t (ql,
•
H E1 - - E?v») . CO (
A matrix analogue of AAK theorem (dim E.«oo, dim Eg, < 00 ) is contained in a very' interesting paper by J.Ball and W.Helton [4J. The most general case has recently been proved by S.R.Treil (Leningrad), [)5]. To begin with, we state a generalization of the Kroneoker theorem. Theorem (Treil. 1984). Let ~eL co( E,,-E t ) , Ek (K = 1, t ) be Hilbert spaces. Then H<.p is a finite rank operator iff ~ ql is a rational funotion of the form
where A'J1.e IV , C11., I< e ( E1<00,0< rank C,-a,k"", • Moreover,
and
<j.) C'H., k
•
,}+
k ~ k'H.
ing induotive rule (in } ):
E~
)
and rank
c,,,,,I«
are defined by the follow-
96
N. K. NIKOL'SKII
1>
Bote, that in the scalar case (dim E~= d~mE~ (Kl
we have Cr., kH- ..r. = C~, k", and the general formula gives rank Hill = ~ k1\. = dfi P_ 4> , as it is prescribed by the Kronecker theorem. Now, put for the convenience d,e.g.. 'V = "f..tltl1!k H'" for every --+- E~) -valued rational function '" and denote by 1?w the set of all \II I S with deg l/f ~ 11, • 4A.
(s
Ek
Theorem (Treil. 1984). Let 4 e Lco (E,,- E~ ) , (k = 1,2) be Hilbert spaces. Then
~11r
=
= ~i~ t
tt1.;f{
IIHqI-H'IIII:
(4)) lt~ + Hco (
"f,().,~k I-/'f~u.}=
E., -- E~ ) ) .
Both inf and dist are attained. Outline of the proof. Fix an integer"t\l and put
wi th ~~= 6'H. ( H<9) • Further, consider joint operator
A=
~~. I - H;
and note that %
1-1 <9
,'W ~
0 ,
the selfad-
(1)
is simply the set of all non-nega-
tive elements of H~(E1) with respect scalar product (A f, 9-) :
to (indefinite)
HA-PLlTZ OPERATORS
97
(2)
The following remark is the main observation of the proof: if t is the shift operator, Sf= 1.f ,then
and hence
fix
eX.
(3)
Let pT, P be the spectral projections onto the non-negative and the negative part of the spectrum of A ,respectively. By property (',) of ~ -numbers, rank p-~ n • Now we are in a position to apply the following lemma. Lemma. Let 1t be a Hilbert space, A be a self± adjoint operator on ';It and P have the same meaning as before. Let further S be one more operator on)t • Suppose that 1) the negative part of A (i.e. A I P-'Jf, ) is invertible; 2) inclusion (3) holds for :x; from (2); 3)
pT S P-E 1'00 .
!hen there exists a closed subspace to of 'Jf; such that i) fo c 'X , ii) 6, is a maximal (by inclusion) linear subspace of:X , iii) $ ~ c & • Replacing condition 1) by a more restrictive one A~=I , one obtains a result due to I.Iohvidov.
98
N. K. NIKOL'SKII
This result was derived in [14] from the Ky Fan's fixed point theorem [15]. Indeed, small variation of Iohvidov's arguments leads to the Lemma stated above. Let us return to the proof of the theorem. Since conditions 1)-) are certainly satisfied, we can consider a subspace & produced by Lemma. It is easy to see that the maximality of ~ implies
On the other hand f, is an invariant subspace of the shift operator S , and together with the property codim ~ ~ 111 this gives a representation &= BH ~(E~) with B a Blaschke-Potapov product of the form N
B= where {,>, K
n
k=1
(~,\ '1t"K +
I(
= 1~A-1. 1(1.
, I ~k I "I<
,I AI< 1< 1
and '11: K are or-
thogonal projections on E-1. (In fact, it follows ~ ~ 4 clearly from ($ - AIH=@, ~e H ( E4 ) that 1= -ed-A~'- , -e,e. E ,and so Ke'l- (s'~JI) I f/ = '1rE1 , ( {- 1. )-1 for a projection 'JL on E1 • We have
J
C, c{ '1r E,,' (
i-X ~)-~ jl=
{
5e: I-f~( E1) : 'Jr"i(A)=O}=('J[gt( 1-1t» H~E4) des-
and this means that the subspace Gf is an
-
<']t'g,\+( 1-11:'» b
S -invariant
subspace of Hf.( E.,) ,codim G..,< The induction finishes the proof.) is non-negative on G ,1/ H~ I GII ~ ~a.
< co-dim G , Since A and hence II 2.q> I BJ.I--( E1)1I=11 e.
'
HA-PLITZ OPERATORS
99
II e.cpBII=11
By the Nehari-Page theorem
and the result follows. 2.
H'If ~ n
•
INvERSE SPECTRAL PROBLEMS
The problems consist in a Hilbert spaee characterization of Hankel operators and of the moduli of Hankel operators. The last of these problems was raised by,S.Hruseev and V.Peller in their contribution [13] to the book [16]. A partial solution has been proposed by S.Treil in [36]. To begin with, let us state an important particular case of the problem. Question 1. Given a non-increasing sequence { ~ 'tf, } n ~ 0 of non-negs tHi ve numbers 'H does there exist a Hankel operator I.f' with t>1\-(
,
? In view of the AAK theorem this question is
equivalent to the following one. Question 1'. Given the same t ~~ } 11-?> 0 does there exist an analytic B MO function <.p (i.e. q> e B M0A ) such that the best rational BM0 -approximations for
)
100
N. K. NIKOL'SKII
It is worth mentioning that replacing here the rational functions by polynomials and the BM0 norm by the uniform one we arrive at the wellknown inverse problem on polynomial approximations which was solved long ago by S.Bernstein. One more reason in favour of the importance of questions 1 &1' is their close relations to the problem o~ the existence of Gaussian processes with prescribed mixing properties. Referring for all the details to ~3J we extract from it only a couple of words on the mentioned relations. Namely, consider a completely nondeterministic stationary Gaussian process 'J(i~d u.e'lt It is known that many mixing properties of the process (like different kinds of ergodicity) can be expressed in terms of the so-called angle operator (PQP){/~ where P, Q are orthogonal projections outo the past span (X-u. : n < 0) and the future Ii' = =span (:lIia : % ~ 0) of the process (see [13]). Further if we write down the spectral measure .f of t Xtr,J 11,eZ in the form dJ'At=lhl~dm with it, an outer 1-/ 2 function (this is possible in the completely non-deterministic case), then the following lemma reduces some probabilistic problems to the Hankel operators (see [13J ).
1
Lemma. The restrictionPQPI fi1 is unitarily equivalent to Ii~ li tt with t(.,=1i,;11, • The following general question (also related to probabilistic problems) was also raised in 03]. Question 2. For which non-negative selfadjo~t operator A does there exist a Hankel operator Htp with I Htp I unitarily equivalent to A ?
HA-PLITZ OPERATORS
101
Now, it is very natural to add two more questions to these two. Question 3. How to describe the set of all ~ from the last Question if this set is non-empty? Question 4. Which operators on Hilbert space have Hankel matrix with respect to some orthonormal basis? If we replace here "basis" by "pair of orthonormal bases" (maybe different in the departure and arrival spaces), we come back to Question 2. A very interesting particular case of Question 4 is raised by S. Power [26] in his contribution to [J 6]: does there exist a non-zero Hankel operator with zero spectrum? To conclude this preliminary discussion we quote two interesting remarks concerning Questions 1,2.
Remark 1 (see [13]). There exist two natural and easy constraints for an operator to be unitarily equivalent to the modulus of a Hankel operator: a) always 0 e: ~pec ( J H~ I ) (because U:" III-t~ l n.1I~ =
t~m II ~ t)1, q> I ~ = 0 ); and b) either Ke."tI I Htjll = { 01, or aim Ke1- I H
=
and hence KfIV I Htfl = Q if KelltIHtjll:::;={CD}
QH.t. ).
with
an inner function
102
N. K. NIKOL'SKII
By the way t in [13] the authors put forward an interesting conjecture (in case of compact operators) that these properties (0 e !I?eG A and either Kl'IVA = UD} or di-m KVI! A= 00 ) are also sufficient for a non-negative operator A to be unitarily equivalent to I Wcp I for a lp e L00 • Two theorems below support the conjecture. Remark 2 (S.Janson). If I ~
~1f1 'tIC)
in P_
have (1-1'1' 1,
and from. Hankel equation (4) we tIC Hcp 1)=( ~1f11, l"'e.. l J
for 11, , I< ~ 0 • Let ~ ,,1 = P-.q> = (J,111. ....1 i ",-t-1 + ... with 0.."'+1*0 t m, ~ 0 • The above identities yield
for 14?; 0 wi th I e,h\+1
t
K~
I=
,and hence HlI' 1 = ~tK+1 l. tIII+ •.. I Q.m .....1 I • Putting / ( = m we obtain 0
and the result follows. Now, let us turn to some recent results concerning Questions 1 and 2 (see D6] for details). Theorem (S. Treil). Let {:JIi tll } \1. ~ 1 be a bounded sequence of distinct positive numbers. Then there exists a Hankel operator HIf such that
H
~
'H.~f
:tt1, (f, itt.)
9-tl., f e. 1-1 ~
HA-PLITZ OPERATORS
where t*~J in Ht and
103
-
and 1~~} are orthonormal sequences 1-1: respectively. Hence
and in the case :t~ + we have ~J1. = ~J1. ( H~ ), n?; 0 . Moreover, the case 0 .. ctO'~ {:t,.} l1-~ 1 forces dim KtJI.. H~ = 00 , but in the case Oe ct06{xKl11.~1 both possibilities ~.{.m KVtr H~ = 00 and KVL H4'= {O~ can be realized. Theorem (S. Treil! V. Vas:yun1n). If { ~.." 1'K-?> 0 is an arbitrary non-increasing sequence of non-negative numbers then there ensts a Hankel operator \-I ~ wi th ~lII
H-L f = 1-)..
P_ ... 1 1 (f-f
where Ae ID , fe H~
~-
,which implies
,
N. K. NIKOL'SKII
104
'2, ~/'N f Z. 1. -1 with .fAK (1-IAICI) (1-A"tf, ~'\I< (1-1'\1(1 ttl-AK) . If the systems {t).l(} ,{ if;.." J were orthogonal, we would have ~~et; ( 1HI) = {O, X,p." , X 11 J • However, these systems are never orthogonal, but for a sufficiently sparse set {AKJ:=~ they are aIm 0 s t o r tho g 0 n a 1, and then :,~ec( IHI> is very close to • Next, by 1, ... , X-H-J a small variation of the coefficients XI<--+ x,*1( we can make the spectrum spec (I HI) to be equal to { 0, X'P'" , xn, J • The proof can be finished by passing to the limit (tt-- (0) •
to,:t
More details. Fix an infinite Blaschke set A , AcID , r:; (1-1 AI) < 00 • The desired A~' s will be ,\e t\ chosen from A according to the following inductive rules. 1. Let A1 be an arbitrary point, A~ e. 1\ and
1-lk,l) (I-A 1 )-f.
2. Induction predicate. Let Ai"'" and suppose that
have the following properties for m ~ a. an appropriate choice of T=,(tn) (+(tnJ
AI1,
vi , .:.,
€
A
and for t(1'K) 111..
€
m . r <m) Jt l m' {a. with some orth onorma1 f ami1 loes 1t J" ( (I<m) J1
JR.""':
"
HA-PLlTZ OPERATORS
for 1 ~ I< c)
~
105
'111-1 ,
~ ~
111
~
W ;
~et { 3:>j({TJ}~ at'S- J-,1<-1
=1= 0
for
T-Tc""where ~k
t:le"
A
q> tI.+1 wi th
Ae
=
"PK
er) +
:It +1.+ 1 (1;-
t "+1)
~-l AI'" l. - A ).
/\ \ \)'1"'"
Al1. 1 ,and let I-{ = ~ ~"
tl+1
= H<.p,,{T' ;- ~nT1 (1 + ttl,+1 )(., f ~) ~ ~ .
Nerl, consi-
der the orthogonal projections I A ' ~ ~ of SA onto the subspaces l-J9" e span (f~M.): 1 ~ I< ~ 'K<) , II~
e
(M.)
,t>.
•
1~ I< ~ M.) respectJ. vely. It is easy to see that they both are the model spaces ( ~te Bn HZ. 9" ~ ft and 1-'- e Bl1 Hwith B", = n ~A ) and hence M_
'!>flQ,n.(~K ~
k-1
I<
So, the both differences f). - -5 A ' ~). - ~ A tend to zero as 1- 1 • Consider now an auxiliary opera tor Hn....1 ,
'h
106
N. K. NIKOL'SKII
namely the orthogonal sum
where
5-11.,.1 = i" / II~ /I
1>p.ec (I
HtH-1 \) = { 0,
'
~1t+1 = 9,). / II g,dl
Xp'" ,
•
Obviously,
oc,a , oc.nt".f <1+ tn+1 )
J
for l' = TCI1 J. But the operators I H~+1 I and I HAI are as close to each other as we desire ifJAI is close enough to 1. Thus for such I AI's the spectrum of I HA I is simple, the eigenfunctions of I H"'1"1 I are close to those of I H~ I and the Jaoobians
and
a~ HA) K(
at j
},tl.t"1
I
,b 1<=1
are close to each other too. By induction hypothesis the first of them is not zero and so is the second. Hence, by the inverse-transform theorem for every A with small 1- I AI there exists a vector l' (1H1J in a neighbourhood of (T (k), 0) such that all the properties a)-c) are satisfied with 11- replaced by a-t 1 • C aut ion : Be careful applying the inverse-
HA-PLITZ OPERATORS
107
transto:rm theorem. In tact, what we need here is a specitication ot this theorem adapted to a family of mappings (and an estimation ot the domain ot inverse transto:rm not depending on the parameter); see [36J. 4. If a sequence t q)nl t\. ~ 1 ,Ql,.=
ti;r
5- ~"')
,9-1( = U:" ~ ~1f,)
•
It is clear
that t :fk } K ~ 1 and {~k ~ I(~ 1 are orthonormal and ~ that H is a Hankel operator. Moreover, /(eJL.H:::J BH since KeN ~ lPK:::J B'\1, ~ ~~ B~ 2- tor every 11- , where
B
U ~).
1 I( • Henc e, dim I< tilt H = 00 • Thus the proof is finished tor the case d~m Kt", H, = 00. As to the existence ot H
Remark. Assuming all A'S to be real we produce
'"' 4
a non-negative Hankel operator since tor c,pH.= 1::
1(-1
tor every f e
H2,
1-1\1<,
•
A few words on the proof ot the second theorem. The main idea (due to V. Vas7UUin) is to use to:rmula (5) § 0 to enJ.arge a gi.Tell tbi. te Bankel
108
N. K. NIKOL'SKll
spectrum spec (I HI)= { 0, Xp
''',
X,..,}
(counted with
mul tiplici ties) by one point :lIi11-+ f = :Ie 11"t,t, = , .. =X11-+1< (with prescribed multiplicity), and then to pass to the limit. We start with adding such a point f r o m the rig h t of a given spectrum. Let ~11t < < ~11t+1 and let k ~ 1 be a prescribed multiplicity of oX "'H 1 • Consider a new Hankel operator, namely 1-1' ~ ~ • !hen, by the Behari theorem ~'= Wi with M1 Ilf1\oo \1 !-IfI <; 1 • Since rank Hf =11, ~f is a rational function, ae~ P- f = 11.,there exists a Blaschke product B-1 , ~e~ B-1 = r1t suoh that B", i ~e;f = ~e;WlO) \I~ II 00 < 1 • Now, one can write 1-1' ~B-1 • In fact, we can replace ~ in the last fonnula ~y an appropriate Blaschke product B~ with deg B~= = deg B" + k • To see this it is sufficient to apply the famous Schur theorem to the ~terpolation nodes A", ... , A11. (the zeroes of B-1 ), An-t~ ) ... , A1-1-+1<' (some new points in the unit disc different from the previous ones). By the Schur theorem (see, e.g., [17J) there exists a Blaschke product B~ , deg Bot = = n+ K which coincides with 9- for A= A~, 1 ~ 1- ~ ~ 11+ K (multiplicities are counted in the usual way). In particular, B1(~-B~)€~oo and henceHB4g.=HBfB~. Now,it is easy to see that 1
'f:
ml
HA-PLITZ OPERA TORS
109
A detail of the proof. In fact, for the proof of Treil-Vasyunin's theorem we need to add a new eigenvalue ~~T1 (with prescribed multiplicity) f r o m the 1 eft of a given spectrum ~ ~ ~ ~ 1'\-1 ~ ... ~ ~ o. So, to apply the previous arguments we must know how to turn our spectrum into ~~1 ~ :'"1 1 ~ ••• ~ ~ -;;, preserving its Hankel character. This aim can be achieved by the following simple but a little bit unexpected corollary of the Sarason interpolation theorem (see [17J). Roughly speaking this corollary (Lemma below) means that up to isometric factors the left inverse of the restriction of a Hankel operator ~ to (KVt, H)l is a Hankel operator too. Lemma. Let ~ e ~ 00 and let 6) be an inner function relatively prime with the inner part of ;f- • If the restriction ~G;f-I K , K des ( Ke'V H~f / is left invertible, then there exists a function ~e: ~ 00 such that (QHe.t I K)-1 = 9~Q9-1 K • Proof. It follows from formula (6) §O that Q ~9!- = f( M&) PG Taking into account the equali ty K= KG we conclude that 0 1- ~p,ev (i ( M(y» • By [;31] there exists a function ~,~€ ~ 00 such that HMef1= ~(MG) and one more application of (6) §O completes the proof: (~~&:f/ kG f'= ~(Me) =
=@H e, IKe' •
So, having a Hankel operator H= H
110
N. K. NIKOL'SKII
lemma ~p-ee (19Hn I KI) = {~-1 ~-f} and spec (I Hi1\" /) = -1 -1 ~§ 0 , ... , "' • "9-lO,~o,"" ~t1.} • Now we are in a posit~on to perform the first step of the proof to produce a new Hankel operator H' with ~p.ec ( / H'/ ) =
_,
_ (
-f
-1
-1
-1}
-10'~0'···'~11.,~t1. ... 1'·"'~1HI(
-1
,
-1
-1
~1S-<~"'+"==·"=~K"'K·A repeat-
ed application of the leDlDlS yields
J.I tft
'Tl\.-t1
~fLe.C(1 H~k~ IJ={ 0, ~n"'K'"'' ~",-t,,, ~1'1.)''')
60
wi th
1.
The passage to the limit can be justified.
3.
TOEPLITZ
O~ERATORS
AND EXPONENTIAL BASES
One more problem of the Ha-plitz spectral theory is the problem of invertibility of the Toeplitz operators. Formally speaking, this problem is solved (for scalar symbols) by the well-known Widom-Devinatz theorem:
where
M~
lP
is a measurable argument of
1m
(
HA-PLITZ OPERATORS
111
One of the aims of the section is to explain connections of the problem of invertibility to some geometric properties of reproducing kernels of the model spaces ke • These connections are rooted in formula (7) gO. The exposition is based on [1a],[17J, [12]. Let us start with making use of three known 00 ' theorems on ~ -functions to reduce the problem of invertibility of the Toeplitz operators to a very special case. These three theorems are the following. Theorem (uler, [3]). Let If eo L00 ( T) • Then there exist a Blaschke product {, and a function :€ H00 -t- C('U') such that fi' = {, f • Theorem (Wolff, Da]). Let .fe~ooT C, If /=1 a. e. on 'lI' • Then there exist an inner function '\} and a function u.., ",6 QC de-it ( HOI1+ C) n ( WlO + C) such that &= ~ u, • Theorem (Sarason, [29J). Let Us Q C , 1,",1=1 a. e. on 'lI'. Then there exist an integer 1+ and real functio~s ~,f, € C <"[') such that u=l)1.e 1,cat-1J where ~ stands for the harmonic conjugation. From these theorems we can easily deduce the following corollary. Corollan 1.
Let
ql6 L00('11'),
e~ji11.;f I <¥ I > 0 . '1l'
!rhen there exist inner functions e and.B , an outer function ~ and a continuous function ~ such that
112
N. K. NIKOL'SKII
and 1'19
is invertible or Predbolm iff so is
Proof. Let ~ = \jI rr where function with 1'\1\ = Iqll OD T lar. Thus, by Axler's theorem 00
fe H
+
'" and
'res •
is an outer Ii' is unimodu-
C,
by Wolff's theorem
by Sarason's theorem
Hence,
(To construct B l. ." either to Ib ~
).
a
and from" and'" we attach or to "" depending on the sign of
rr,.
By formula (2) ~O, 'r(jl=1'B9h. '1'1 and is an invertible operator since 9- is an outer function (9-e. Heo) with e~~rrl'H.f Ifl ~ 0 • As to the function h , i t is a so-called multiplier preserving the invertibility, i.e. the function such that for every Ve. Loo t 'rv is invertible iff so 1s 'r v1t. •
HA-PLITZ OPERATORS
113
It is easy to see (e.g., from the Widom-Devinatz cri terion) that every It, of the form i1-= e,c. , c€ C<'ll') possesses this property. It is also quite obvious that a multiplier preserving the invertibility preserves the Fredholmness too. • Incidentally, the above assertion on multipliers is an immediate consequence of the following criterion obtained in [34] and independently in [18]. Theorem 2. Let}A'eLoo ('1l') • The following are equivalent. 1. For 4>e L00< '11') ,rr~}A is invertible, if so is T~ . 2.
(.~~itlf Ifll > 'Ir
+ C.to-~Loo
3.
0
and
)4=/ }AI-e ill , o.e'Ret:o+
Rf.. H
'rJAil(
00.
is invertible for every lX. > 0 •
So, the corollary derived above reduces the general problem of invertibility of a Toeplitz operator <9 to a special case of symbols of the form 41= Q~ G~ , 81 and g~ being inner functions. The properties of the latter operators are closely connected with some geometric properties of the model subspaces and their reproducing kernels.
rr
Corollary 3.
Let
91
,
e~ be inner functions.
Then 1) r9i~~ is left invertibe iff the restriction Pe I Ke is an isomorphism (onto its range). 1 ~ 2) ",l\Vi l\ is invert 1 ble 1ff so is Pe I K" . Qt 1 Q~
114
N. K. NIKOL'SKIJ
Proof. Immediate from (7)
§o.
Corollary 4. Let B be a Blasohke produot with zero set 6" (i.e. B= IT ~'" 6"C1D, !: (1-1~1><
e
and
Ae:tr
AE:~
00
)
be an inner :f\motion. Then
T9B
is left invertible 1ff
lit /6"= Kel
8' ,
i.e., iff the following interpolation problem is solvable
r
Proof. QB is left invertible iff 'f~ ='P is onto, and henoe (from formula (7) §O) iff PBKe-KB • But it is obvious that the last equality is equivalent to interpolatory assertion stated ab~ve • •
'res
eB
Corollary 5. Let Q and B be as before. If is left invertible then 1) the reproduoing kernels ~ kQ (A, . ) j Ae 5" '
e
1- Q (A) form a topologioally free (or 1- At minimal) family of funotions; k (A')
G,
=
The proof is immediate from Cor.) if we take 1 into aooount an obvious formula(kQ(A,.)=P6(1-A~f, Ae ID ), a known defini tion (a family of veotors { x A } AE5 6' is oalled minimal if ~JA ¢ ~p.a. no (X" : ,\ e 6", A=1= JJ.) , 't/ M e ~ ) and an obvious and known fa.ot
HA-PUTZ OPERATORS
115
the system {<1-1'lri j >.6
0"
is minimal under the
Blaschke condition ~ (1-1 A\) . .: :.
00 ).
•
AE6'"
Corollary 6. codim span
=d,,:w{, Ken,treB ' = Ke iff reB
(k e<)..,.) : A€
and, in particular, span
~) =
(kacA,.): A,€ 6")=
is an injection.
This follows from formula (7) §O and from
span
«( 1- 1"1;)-1 :
).. E 6")
= K.s .•
It is not hard to see that Corollaries 3-6 contain as a very particular oase refor.mulations of some interesting properties of exponentials in terms of Toeplitz'operators. To make things more clear let us write down some known formulae for transformations of
a" ~ 0
I~ mt.
LJ (
u.).
In these formulae
; KtV =
KeQ, ,p(1.=
II
~Il. =
P~Q.,'
-toc,p a 1,+1 'i\-1 '
&a;- the space of
entire functions of exponential type to L1 (i) ,
~~
belonging
x elR , ! being defined on 'Il' rier transf0:rm.
,and g:
stands for Fou-
N. K. NIKOL'SKII
116
I Z, rm LJ ( u)
is unitarily map~ ped by U into
L~
is unitarily map~ ped by ~ into
II)
( ~)
wt______• Ht(Jm ~70> (.(eJ H~-------" L~(O,oo) ea.H~--------"'''~ ~1.Q~ H~---------l~-L ?(a,oo)
K~ H~ 9Q, Ht_____ • ei~x, GQ,/~ _ _ _ _ _ _--'lI~- Lteo, (1,)
pQ. -----------------~,.,- (.f-t(o,a) f)
(up to unessential constants;
kQQ (A,
.)
,.,
Thus, ru transforms corollaries 3-6 following propositions.
1.16 B
1-(0,0)
e, iJloc,
into the
is left invertible iff there exists a constan% c ~ 0 such that Jool :f/?"~ c for 15-/1"
te
~p'lUt. L~
2.TQoB
fe f<jA)
H; =
J4.
~)I:Jr,
(e
0
is left invertible iff for every
there exists an • Q
e 1. T}4
0
: jJG7: ) •
9<jL),
pe.rr;.
ifE GQ,/IJ.
such that
HA-PLlTZ OPERATORS
3. It
117
rga. B
is lett invertible then the tami~ 1y { e1.JA:r. : pe'r} is minimal in L (0, a.) • .
It is appropriate to note here that,tor example, the last proposition connects the two well-known pr~b1ems, namely, the problem of the description 01' exponential families with the completeness property on an interval ( ~P.Cl.f\.l,,~CO,tl.) (e 1}4':IC: perr: ) = Lt (0, Q,» , and the problem 01' the calculation of the point speotrum 01' the Toep1itz operators. See tor details [271 [32] (for the first problem), [6] (for the second). It would be very tempting to find a solution of the completeness problem based on knowledge of the point spectrum of Tea.B • More deep relations between the Toeplitz operators and the exponentials families can be establiShed for a special but important case, namely, fQr the tamilies \ e 1.)A:Ki : p € 'r; } wi th sparse set of frequenoies. To demonstrate these relations let us recall some known notions and results about the bases of rational. functions {(1-1~f1 : A6 ~ 1 in U~ • Under the transfo1"lDs U and ru they can be unitarily transferred into J.I~ and Lt(o,oo) ,respectively. ]'or details see, tor example, [17]. By detinition, a family F'= { !-). : A6 6' 1 of vectors of a Hilbert space H forms a Riesz (unconditional) basis in its span i t the following (approximate Parseval) inequalities hold
cI: I(J.AI~II.f"IIi.~11 E a,At~ IJt,~ CE 1a.>.Il(,U5,,/lt. Ae6'
A... f'
.hH'
118
N. K. NIKOL'SKII
for some (i, C.,. 0 and for any finite numerical family to.,>. J • It is easy to see that the family fi' , being a basis, is minimal, and even uniformly minimal; the latter means
It is not hard to check that for a family of repro. -1 ducing kernels f,). = ( 1- A~) ,A € 6" , we have
Corollary 7. If t(1_X~,-f : basis in its span Kg ,then
Ae6" I is a Riesz
( (C) => Riesz basis property) is also true, but this is a much deeper result which is essentially due to L.Carleson. There are some useful and transparent reformulations of the Carleson condition (C) in terms of the distribution of ~ in the disc; see [17] for a discussion. Thus, the problem of exponential bases { e iJA:X- : ,ftA(,ii,'t"} :Ln L~(O, (0) is solved. However, the situation is not so simple for L~ spaces on finite In fact, the converse conclusion
119
HA-PLITZ OPERATORS
intervals of the real axis. W10g we can deal with t LeO, Q,) , (;t > 0 • So, the problems are: to describe all the Riesz bases of exponentials t tt~~ ~ JA e ~ J , 'r c t in the whole L~ ( 0, at) as well as in span Lt(o,Q.) (~ip~ ~; e:. '!;')
•
The method of
Toep1itz operators described below works only in the case of the frequencies 'r contained in a ha1fplane Jm ~ 7 C ,or ~m %<. C ( CG /R. ) • Since the mappings f(~)- e'--'1~ 1-(Ot), ;rex)-f((;t-x) are isomorphisms of L~O,Q,) onto itself, we can restrict ourselves for definitness to the case 'V c
lJml>c;;.OJ.
We start from a theorem on general properties of bases of reproducing kernels. In what follows this resu;L t is applied with S= ~flI' ct. > 0 . Theorem 8. Let 9 be an inner function and 6"c /D • If the family { ke (ft., ) ..:. Ae ()' } is a Riesz basis in its span, then 1(1- A1, )-f : ,.\ e 6' J is a Riesz basis too (or, equivalently, the condition (0) is satisfied). Referring for the proof to [171 we give here an easy and short reasoning which is sufficient for our main particular case 9 = 9Q,' £t > o. Namely, under the additional hypothesis ~~
Ae
~
I E) ( }.,) I <
1
(1)
the proof turns into an obvious remark of geometric nature. On the otJler hand, for exponentials
120
N. K. NIKOL'SKII
( i)x-}
(9= (9Q,) tn;f { j-m, JA : .p e. if: } > 0
1. t
)J.err:
condition (1) means which does not lead to
loss of generality. So, (1) implies" Pe (1- 1 'l )-1 II~ =
(1-/ etAl It)( 1-/ A/t f 1~ C011~t <1-/ AI~l-f Cbn4t I (1-;;1I rf I ~ with const independent of AS ~ • The desired Carleson condition follows from the uniform minima1ity of the family t (1-11, f': Ae. ~ J while the last property can be obtained if to put T== Pe in the following e1ementar.1 lemma.
=
Lemma 9. Let { i-A : AC 6" J tors in a Banach space X and rator with the property
be a family of vecbe a linear ope-
r
If the family {T fA : Ae ~ 1 is uniformly minimal then {fA : Ae 6" j is uniformly minimal too.
rf), r\ ~ p,ll11- er f.,M : ~ =FA » = =111,\ II-liT 5"). 11-1 ~i~ t ( r eTA ~ of A II-I, ~patL (rr 5-p :I-FA » ~ Proof.
dist ( T 5,\
II
~ " l' ~ ·lloT),!!-!! r .f). rf ~i~ t (! ,"I .f), rf, 6 pan ( of}1 : f4=FA» _ • How we are in a position to join all the previous remarks into the whole picture describing the interplay ot the invertibility of To.plitz operators
HA-PLITZ OPERATORS
121
reB
and the reproducing bases problem. As a background 0D? can consider Corollary 3 and the following simple but very important remark by B.S.Pavlov (see [12]): if (2) holds and if both the families { 5,\ J and {rr J,\ J are Riesz bases, then 'T' is an isomorphism. Theorem 10. Let 9 be an inner function and B be a Blaschke product with the zero set ~. The following are equivalent. 1. rS'B is left invertible and 6"e (C) • 2. The family { ke (A,') : Ae 6' } forms a Riesz basis in its span and
3. dist
sup { I 9d)\: )..e ~ }
(e B , H00) < 1
and
<
1.
6"e (C) •
One can easily modify this theorem to consider. invertibility 01'198 and bases of reproducing kernels in the whole space Ke • Then an additional condition appears in 3, namely, the condition di~t (QB, It oo ) <1 • It is curious that both inequalities d.,i,~ t (GB, HQI><1 J rU~t ( 9B) H00) <. 1 imply that M~t <eB, Hcc ) = ~.(,~t (SB, ~ 00 ) • In fact, following V.Peller we can see from formula (5)§0 that the operators It'\.t- and ~u; are unitarily equivalent if Iul= 1 a.e. and Ken., 'r'\A..=Ke/J/r.u: = {@}, and hence
H:
H:
di ~t ( u" H00 ) = di~ t
(1A.,
H00)
•
In the conclusion let us notice that with the help of the explained connect~ons between the Toeplitz operators and the bases of exponentials (for
122
N. K. NIKOL'SKII
8 = SQ,
it is not hard to prove again all known facts on Riesz bases {e, iJA~ : ,.e.~} . in Lt spaces as well as to obtain new ones. We refer to [12] for all details including some historical remarks. It is also possible to restate all the results on Riesz bases of reproducing kernels (of exponentials, in particular) in terms of solvability of free interpolation problems on the model space Ke (on the space f; a. of entire functions in the case = 9 Q., ) .
e
4.
SPECTRAL MULTIPLICITY PROBLEM
There are some classes of operators for which the multiplicity of spectrum plays an essential role. The first of them is the class of normal operators on Hilbert space. For a normal operator the multiplicity has a local meaning and is essentially a unique unitary invariant of the operator. Namely, if a normal operator, say N ,is represented as the mul tiplication by independent variable, fO,) - - - ~f(l) ,acting on a direct integral
J ~fJ.e.c.
G)
H(~) ~H ~)
( N)
with respect to a scalar spectral measure, say ~ then one can set .fN(~)= ~im~(~)7 ~e6fUc(N) . Hence the local multiplicity JAN (0) is an s-a.e. defined function. In the case of a pure point spectrum we have simply
123
HA-PLITZ OPERATORS
The same formula can serve as a definition of a local multiplicity of an operatar on finite dimensional space, but in this case it loses the property to be the unique unitary (and even similarity) invariant. However, it is still useful. Moreover, a more rough numerical characteristic is still useful, namely (1)
called the (global) multiplicity of the spectrum. For the both classes of operators mentioned above the last characteristic admits the following independent description. Lemma 1.
fkr =
~in {card C: span
(r"'" c: n ~ 0 )
(2 )
is the whole space} . In other words foT is the minimal cardinality of ~ -cyclic sets. For finite dimensional spaces the proof follows from standard (in Jordan model) considerations. For a normal operator N with simple spectrum, i.e. for foN = f , lemma 1 was proved in [5]. For .fo N > 1 a similar reasoning is applicable. Since the right hand side of (2) makes sense for the most general operator ,one can define the multiplicity of the spectrum fo~ by formula (2). As to the local multipliCity there is no reasonable way to define it for general operators. It is worth mentioning that for an operator '1' with "rich"
r
124
N. K. NIKOL'SKII
point spectrum one can regard the quantity dim Kf/t, ~ T- AI )*, Ae :,p,ec. (1') as a candidate for a local multiplicity of l' rather than the quantity dim KeIL ( ~I) , Ae; 6p,ec (1') • In particular, some results of [20] point to necessity of such alteration as well as the following elementary lemma berrowed from [20J does.
rr-
Lemma 2. For every operator
Proof. Let spec ('r) and C c
t Km T*}l.,
'T'
A= 0
be an a.rbi trary point of be a cyclic set. Then 'P 11 C c t1,?> 1 , and hence rAM,a, c?> dim ~P'(M1 C~
~ c,o.~i fI1, ~-p.(M1 ( l' 11 C:
J1.
~ 1) ~
d1. m Ken, rr *" . •
Let us turn to the Toeplitz operators. One can expect to develop a notion of the local multiplicity starting from some D.Clark's results [7],[81. To be more specific, let us consider (for example) <.pe. G('R' ) and A¢ t~Hp,e.r,rrtf = lP (If) • Set fl~ (A) =
=dimKI'JI,('l'Ij>-AI)~ P~(A)=dimKtJL('r4'-AI). This pair of functions can playa role of·the local multiplicity (as a similarity invariant of an operator) on some classes of ~tf 's. Indeed, one of the theorems of [7] says that if'
HA-PLITZ OPERATORS
125
ing Up" condition can be omitted if one completes ± the system of invariants Jk~ by one more (normal like) local multiplicity function, say }t~ (~) ,defined for A€ c.p ('I'> where <.p "backs up". Again, rr'" rn ± ± and 1'1' turn. out to be similar iff .fo \II === P 'II and ~~ == }t~ • In fact, in [7], [8] it is proved that under some (restrictive) assumptions like those mentioned above the operator ~~ is similar to an orthogonal sum of operators of the form (3)
*i
(resp. ~~ ) are conformal mappings from ID onto the positively (resp., negatively) oriented "loops" of ~ (']f') (each sUDUDand is repeated according to the mul tiplicity P~ (A) ,resp. fo~ (~» and the normal operator N comes from the "backing up" points of lP ('ll') • The above preliminary remarks about possible choices of local multiplicity functions can serve as a guide to the following attempts to compute the global multiplicity of the spectrum of some Toeplitz operators. To support these remarks we mention in advance that for "good" rational functions ~ formula (1) turns out to be "almost true" (up to the mentioned advantage of Km (T- >'1)* over I(t/l,('r-Al) ). Namely, Prr = m£t.:Ki }At, d) or JUT = mfl,:t AJ,+ d) + 1 cr ). Cf J.. r~ depending on the degree of t,he influence of the first group of mumnands in (3) on the second one (for details see subsection C below). Formula (3) and the orientation to "good" (not
where
126
N. K. NIKOL'SKII
"backing up") symbols force us to consider the three following cases separately: A. antianalytic symbols; B. analytic symbols; C. multiplicity of the spectra of orthogonal sums. In what follows we wri te P'r~ = 'pcp for brevity. A. ANTIAIULYTIC SYMBOLS This is the simplest case. A dir~ct sum of antianalytic Toeplitz operators, namely ~ ~ T~. ,
he
~}
j=1
~oo
,presented in Clark's formula (3) can be interpreted as a vectorial Toeplitz operator TG-* on the Hardy space
Ht
(fJ
wi th
G-= aia,~ (~1"'" ~K-)
•
More generally, let us consider an arbitrary operator '11. ~e.f H00 Icr*' ere HOOK' (~-c)= ttrtt.
m
t1,
Theorem 3. Let G-e I-l ~ 01,
then
Jkr:.* = 1
• otherwise
•
If
.f'q* =
00
•
Note, that for G= d..j,!l~ (g of, ... , ~ tf, ) (4) means that ~K =1= const for every this is the case for any Clark's sum (3). Outline of the proof. Let P be the ristic (minimal) polynomial of G(O) * • §O we see that p (rr~,," ) = rp(~It") •
p(Gl!) ( 0) = = Ht( t tt)
QD
,
,
we have span l Kelt rp~G-*) :
condi tion I< ,and characteFrom (2) Since K
~ 1) =
and one can apply the following gene-
HA-PLITZ OPERATORS
127
ral proposition from [20]. Lemma 4. Let A be an operator on a Banach space X and let span (Kvv Ak : k ~ 1 ) = X • Then foA= ma,x (1, cl~m KfIV A*).
Now, put A== rp(q~) and note that (4) is equivalent to Km A*= t@l • Moreover, the kernel of an analytic Toeplitz operator (as A~ ) being non trivial is infinite dimensional because it is invariant under the shift. • B. ANALYTIC SYMBOLS
Unfortunately, this case is still incomprehensible, but for some sample classes of symbols (e.g. for a polynomial "in general position") the quantitYJLf can be computed. Let us start with an elementary corollary of Lemma 2. In what follows we denote by ~ 1M. the inner factor of a function e H00 in its Nevanlinna factorization.
*
Corollary 4. Let
e:
H00
•
Then
val e n c y of The last sup is called the ~ and denoted by V~ • The proof of the corollary is immediate since J(e;z. ('T\¥ - AI >*= «
128
N. K. NIKOL'SKII
defining the "e s s e n t i a 1 val e n c y " of (jl , say e. V~' in a natural way (namely, taking into account the boundary values of
ID )
we have ~
e,V~=.2,
e-VI.p(1,M)=
2"", > '\It.
2t1t
, =
•
Vr.p=
1 ,and hence
YI.p (lK)
(tv ~
~
; really, here
has a constant valency (i.e. card tp-1(A) ===. const, Aetp( /D) ) and
However, if
1)
fo
tp
Lemma 5. Let til
be a bounded univalent function in ID , J1 = tp ( ID) and tV be the ha:rmonic measure for.n.. at y> (0) • Then, fl~= 1 i~f there exists a positive function w on J1 such that
and Ce06 9'A =:l I-l 00 (...a.) , the closure being taken in L~ ( '\.Ud. W) • Moreover, if the last inclusi-
HA-PLITZ OPERATORS
129
on holds, the outer function f on 'Jr is cyclic for 7'q> •
with
If/t
'LUo<.p
a.e.
Example, [19J. Let us consider a crescent domain Jl. wi th not "too bad" boundary and a confomal mapping lP from !D onto JL
p
Let for
be harmonic measures for JL , resp. lLo ( -- the bounded component of the complement at some points of J1 , Jl. o • Claim: If there exists a constant C. such that u)
,Wo
(, "..n.. )
on
(6)
then .fq> = ~ ; otherwise jA-q> = 1 • Condi tion (6) means that near the point P the domain JL c is "thiner" than JL • Below we explain only the first of the claimed assertions.The proof of the second part of the Claim is based on Volberg's arguments on mean polynomial approximation [:37]. The inequality P-q> ~ t follows from the possibili ty to write 41 as a cemposi tiol! <.p = 0( 0 J ' where ()( = «l,'l + {,) ~ (0 bviously , foot. ~ Z ) and J is a conformal mapping of ID onto a Jordan domain (unclenched crescent, JA, = i ) and from the following proposition, [19J.
130
N. K. NIKOL'SKII
Lemma 6. If <X, 1£ JAtrl.
Pe Hco, I} I
~
1
then
.p /Xo"
~
. }At} •
It remains to show that if (6) holds then there exists no cyclic function. To this end let us consider a function ~ satisfying conditions (5). By (6), these conditions remain valid if we replace 00 and by Wo and 0 • Let fJ- be an outer function in H~ such that I ~ I't._ WO q>o , lPo being a conformal mapping of ID onto Jl..o • Then, for every
a.n.
a.u.
pe .fA
J I po lPol1"I~I~= J IPI"w~wo" CJ J I pl1. w "-w , 'Il' 8.n.. aJL and hence for every cy~ ct0-6 L'-('lUdw) 9"'A
~G H'" • Putting tJ,= (1 - tori ~o E.ILo we can see that
we have with and • Thus,
Now, we state two recent results due to B.M.Solomyak [33]. Recall that always jJ-cp ~ evcp • Theorem (B. SolonqaJs). Let
Then,
ct~ ID ;
HA-PLlTZ OPERATORS
131
fo~ ~ e VlP +
1.
If the complement (), c:P (
rr
Recall that all such generators have been described by D,Sarason [30].
132
N. K. NIKOL'SKII
c. MULTIPLICITY OF THE SPECTRUM OF ORTHOGONAL SUMS It is obvious that for a pair of Hilbert space operators A and B the following inequalities hold
The left hand inequality turns into the equality if the operators "are independent", e.g. if the polynomially convex hulls of ~cA and ~p.ecB do not intersect: pch spec An pch spec B =
then rAe Tu!' = »ta.~ ( .fA" JAr <:tit ) stands for the complex conjugation,
; ()*
here
p:: Ae
6'*
G'} .
Corollary 9. Let 51, € Hoo ,1 ~ i ~ om, and let 9i be non constant functions from J..l 00 , 1 ~ j ~ 11 • If Ii' = G/,-ia.g ( 1- 1 , ••• , ftK ), G- = =
a,-i, (A,~
(
~,
,.. .,
~
l1. )
and
if
HA-PLITZ OPERATORS
133
Theorem 10. Let a function f ,feH- oo maps ID with a constant valency onto a Jordan domain. Let qi ,~ ~ }~ 11, be non constant and
Then,
frrt ~ T
(lll-
=
JAT!
+ f'T'(i-AA- =
f'Tf + 1 ,
where
G- = tiia.,~ (~1'''' , ~ ~ ) • It is not hard to see that the last theorem and the last corollary give some formulae for the multiplicity of the spectrum for rational Toeplitz operators r~ which are covered by Clark's model (3) without "backing up" summand N • The subject of this section will be discussed in more detail elsewhere.
REFERENCES
[1} Adamyan, V.M.; Arov D.Z. and Krein, M.G. AHaJD!!T~~eCRHe CBo~CTBa nap ~Ta raHRe~eBa onepaTopa
~ oOoOIIteHHaH sa,n;a-qa mYPa-TaKan. - MaTeM. COOPHHR, I97I, 85(I28), ~ I(9), pp.33-73.
[2] Adamyan, V.M.; Arov, D.Z. and Krein, M.G. Bec,ROHe'tlHHe O~o'qHO-raHReJIeBIi MaTP~Illl ~ CBIlSaHHHe C
134
N. K. NIKOL'SKII
HEMM
rrpo6~eMH rrpO~O~eHRH.
CKO~
CCP,
MaTeMaTRKa,
- MSBeCTRH AH ApMHH-
1971, 2-3, pp.87-II2.
[3] Axler, S. Factorization ot L00 tunctiens. - Ann. Math., 1977, 106, pp.567-572. [4] Ball, J.A. and Helton, J.W. A Beurling-Lax theorem tor the Lie group U (111,)1,) which contains most classical interpolation theory. - J.Operator Theory, 1983, 9, pp.107-142. ~J Bram, J. Subnormal operators. - Duke Math.J.,
1955, 22, 1, pp.75-94. ~J Clark, D.N. On the point spectrum of a Toeplitz
operator. - Trans. Amer. Math. Soc. , 1967, 126, 2, pp.251-266. [7] Clark, D.N. On Toeplitz operators with loops. J.Operator Theory, 1980, 4, pp.37-54. [8]Clark, D.N. On Toeplitz operators with loops. II. - J.Operator Theory, 1982, 7, pp.109-123. [9] Douglas, R.G. Banach algebra techniques in operator theory. Academic Press, N.Y., 1972. [1 0] Douglas , R.G. Banach algebra techniques in the theory of Toeplitz operators. - CBMS series, 15, Amer.Math.Soc., Providence, 1973. D1]Helton, J.W. Non-Euclidean functional analysis and electronics. - Bull. Amer. Math. Soc. , 1982, 7,1, pp.1-64. ~2J Brus~ev, S.V.; Nikol'skii, N.K. and Pavlov, B.S.
Unconditional bases of exponentials and of reproducing kernels. - Lect.Notes Math., 864, Springer -Verlag, Berlin - N.Y., 1981, pp.214-335.
HA-PLITZ OPERATORS
135
D3] Hru~cev, S.V. and Peller, V.V. Moduli of Hankel operators, past and future. - Lect.Botes Math., 1043, Springer-Verlag, Berlin - B.Y., 1984, pp.92-97. Q4} Iohvidov, I.S. 06
O~OH ~eMMe K.~aHa, 0606~aID
npHHUHn HenO~BREHOH TO~KH A.H.THxoHOBa. no~.AH CCCP, 1964, 159, 3, pp.501-504. ~eH
05] Ky Fan. Generalization of Tychonoff fixed point theorem. - Math.Annalen, 1961, 142, pp.305-31. D6] Linear and Complex Analysis Problem Book. Lect. Notes Math., 1043, Springer-Verlag, Berlin N. Y., 1984. 07J Nikol'skii, B.K. M3~."HayRa",
~eKUHH
06 onepaTope
C~BHra.
MocRBa,
1980. (English translation including [18J as an Appendix: Treatise on the shift operator. Springer-Verlag, Berlin N.Y., 1984-1985).
B8] Nikol'skii, N.K. OnepaTopH raHR~H n TenHKUa.
CneRTpaHDHaff TeopHH.
AHre6pan~ecKnH no~o~.
ITo-
~e~HHe ~OCTR1KeHM.
~9J
- ITpenpnHTlI ~OMM (~eHHHrp~) P-1-82, P-2-82, P-5-82, pp.1-181. Bikol'skii, N.K. Ha6pocRH R B~n~eHHD KpaTHOCTH cneRTpa OPTOroHaHDHHX cYMM. - 3anncKH HaytlH.cewmapoB ~OMVi (JIeHHHrpa.n;), 1983, 126, pp.15O-158.
~O] Bikol'skii, •• K. and Vasyunin, V.I. Control sub-
spaces of minimal dimension and root vectors. Integral Eq. and Operator Theory, 1983, 0, pp.274-31 1• [21] Page, L.B. Applications of Sz.-Hagy and Foia,
136
N. K. NIKOL'SKII
lifting theorem. - Indiana Univ.Math.J., 1970, 20, pp.135-145. [22J Peetre, J. Hankel operators, rational approximatiGn and allied questions of analysis. - Canadian Math.Soc.Confer.Proc., 1983, 3, 287-332. [?3]
Peller, V. V. OrrepaTopH raHReJIR Macca 1"', E me IIpMOlteHM (PaIUiOHaJIDHaH annpoRcmvra.nrur:, raycCOBCRHe rrpOnecCH, rrpo6HeMa MaltOPanKH onepaTopOB) - MaTeM.c6opHHR, 1980, 113, 4, pp.538-58I.
~4] Peller, V. V.
and Hru~clev, S. V. OnepaTopH raHRe-
npH6JrmteHIDi H CTaIUiOHapHlile raycCOBCKHe nponeCCH. - YcneXH MaTeM.HayK, 1982,3?, I, pp.53-124.
JIJi, H~e
[25] Power, s.C. Hankel operators on Hilbert space. Research Notes in Math., 64, Pitman, Boston London - Melbourne, 1982. [26J Power, S.C. Quasinilpotent Hankel operators. Lect.Notes Math., 1043, Springer-Verlag, Berlin - N.Y., pp.259-261. [27] Redheffer, R.M. Completeness of sets of complex exponentials. - Adv.Math., 1977, 24, pp.1-62. ~8]
Sarason,D.E. Function theory on the unit circle - Notes for lectures at a cont. at Virginia Polytechnic Inst. and State Univ., 1978,preprint.
[29] Sarason,D.E. Algebras of functions on the unit circle. - Bull.Amer.Math.Soc., 1973, 79,pp.286-289. 1)0] Sarason ,D.E. Weak-star generators of Hoo • Pacif.J.Math., 1966, 17, B J, pp.519-528.
HA-PLITZ OPERATORS
137
[31] Sarason,D.E. Generalized interpGlation in Hoo • - Trans. Amer. Math. Soc. , 1967, 127, N 2, pp.179-203.
El1opToroHalIbHHe pa.3JIOJKeHIDI mYHRUMli B p~ 3RcnOHeHT Ha RHTepBaJIax BemecTBeHHoH OC~. - Ycnex~ MaTeM.HaYR, !982, 37. 5, pp.51-95.
[32] Sedleckii, A.M.
° KpaTHOCTH CneRTpa aHaJIHTKtIeCRHX onepaTOpOB TenllHua. CCCP, 1985.
[33] So 1 omyak) B.ll.
~ORJI.AH
° MHOJmTeJIffX, He
[34] Spi tkovskii, I.M.. ~aKTOpH3yeMoCT~.
-
B.7lRfIKIIUOC
Ha
MaTeM.3aMeTRH, i980, 27,
2, pp.291-299. [35] Treil,S.R. BeRTOpHHM BapnaHT TeOpeMH ~aMHHa ApoBa-Kpemra. - ~KIWOHaJI:Dmrn aHaJIH3 Hero IIpHJIOJKeHIDI, I985. ~
~6J
MO,IlJ"JIH onepaTopOB raHRe.mI H 3a,u;a~a B.B.rreHJIepa~C.B.Xp~eBa. - ~ORJI.AH CCCP, 1985. Treil, S.R.
[J7J Volberg,A.L. How to break through a cribed contour. - To appear.
pres-
D8] Wolff,T.H. Two algebras of bounded functioDS. - Duke Math.J., 1982, 49, 2, pp.321-328.
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
Bernt 0ksenda1 Department of Mathematics California Institute of Technology Pasadena, California 91125 USA CONTENTS
SO.
Introduction
§1.
Basic definitions. The Levy theorem: Brownian motion and analytic functions. Brownian path preserving (BPP) functions. Applications: Boundary behaviour and removable singularities for H2(U), for a general open set U C tn.
§2.
Markov processes, generators and Ito diffusions. Pathpreserving functions. A proof of the fundamental operator identity characterizing the path-preserving functions. Corollary: The Levy theorem.
§3.
Applications of the operator characterization: a)
Analytic functions and conditional Brownian motion
b)
A general view: When is a given continuous function an Xt - Yt path-preserving function for some Markov processes Xt , Yt ?
139
r. C. Power fed.}, Operators and Function Theory, 139-162.
140
§O.
B. ¢KSENDAL
INTRODUCTON
It is now well known that there is a close connection between Brownian motion and (classical) harmonic functions. This discovery started with Kakutani' s solution in 1944 [22 1 of the Dirichlet problem by using Brownian motion. Subsequently many other striking connections have been found and they have been extended to general Markov processes and associated harmonic spaces. See e.g. Constantinescu & Cornea [5 1 and Bliedtner & Hansen [ 3 1• Since analytic functions are in particular harmonic, this of course also implies a relation between Brownian motion and analytic functions. But it is not so well known that there is a stronger connection between these concepts than that. This connection is provided by Levy's theorem from 1948 which roughly says that analytic functions map Brownian motion into Brownian motion except for a change in time scale. It is the purpose of these talks to explain, apply and expand this result and show that it is linked to a fundamental operator identity, valid in a much more general situation. This allows one to choose Markov processes suitable for the type of investigation one wishes to perform on a given function. §l.
BROWNIAN MOTION AND ANALYTIC FUNCTIONS
Before we formulate the Levy theorem, let us basic definitions and introduce some notation: 1. 1.
recall
some
Definition
A stochastic process is a family {X t } t>O, where for each t X t is a random variable w + X t (w) definea on a probability space (~,d)p) and with values in ~n. Many important properties of the process {X t } (often denoted simply by X t ) are given by the finite-dimensional distributions: lltl, ... ,tk(G!X ... x Gk) = P[X tl
c Gl,
... , X tk E Gkl
(1.1)
where ti E:. [0,00) and G.are Borel sets in ~n. Conversely, it is a classical resulf due to Kolmogorov that given a family Vt , ..• ,tk of probability measures on ~nk (for all tl, ... ,tk; k=l,2, ... ) satisfying certain consistency conditions there exists a stochastic process {Y t } whose finite-dimensional distributions coincide with v t 1, ... , tk' This may be used to define n-dimensional Brownian motion as follows: p(t,y)
Let If 0 by
<
tl
~
n (27ft) 2 exp(-
t2 ...
<
Ilt 2 ) l
for y, E:.
~n,
t > O.
tk define a measure pO on tl"'" tk
~nk
0.2)
141
STOCHASTIC PROCESSES. INFINITESIMAL GENERATORS AND FUNCTION THEORY
where we use the convention that p(O,y)dy = 80, the unit point mass at O. By Kolmogorov's theorem there exists a stochastic process {Bi} t>O on a probability space (~, :f,po) whose finite-dimensioilal distributions are given by (1.3). This process is called the Brownian motion starting at O. If we replace the measure pO by the measure px defined by ( 1.4)
we get the Brownian motion starting at x. If f E Ll(px) we will let EX denote the expectation of f wrt. pX, Le. EX[£) = ~ f (w)dpX ( w) . One the path
can
prove
t
that
for
almost
all
(a.a.) w t S1
wrt.
pX
<0
is a continuous map from (0,00) into mn. Thus we may identify (a.a.)wE~with the continuous function t ~ Bt(w) and we may adopt the point of view that Brownian motion is just the space C([O, 00), mn) equipped with the measure pX given by (2.3) and (2.4) above. We can now formulate Kakutani' s solution ([ 22]) to the classical Dirichlet problem: Let U c mn be open and let TU = inf{ t
be the f be a Put
0.5)
> 0; Bt f;:. U}
first exit time from U for Brownian motion B. Let t continuous bounded function on dU, the boundary of U.
0.6)
Then f is harmonic in U (this is a consequence of the strong Markov property, see for example [22]) and if au is "reasonable" (e.g. C l ) then lim f(x)
Hy)
for all
y
tau.
(1. 7)
x~y
XEU
-
Note that this formula for f gives the harmonic measure Ax for U is given by
in
particular
that
142
B.0KSENDAL
= PX[B,U
Ax(F)
t
(1. 8)
F] for Borel subsets F of dUo
Next we briefly recall the basic properties of Ito integrals (for details see for example [22]) • For the moment let Bt denote I-dimensional Brownian motion and let I t be the a-algebra generated by { Bs; s < t} for o < t <00. Put 1:00 = 1. Suppose v (t ,w) [0,(0) X rl -+-m. satisfies v(t,w) is ~X I-measurable, where ~ denotes the Borel a-algebra on [0,(0)
(i)
For all t the function w -+
(ii)
EX[6v2(s,W)dS]
(iii)
< 00
for all t
>0
Then we can define for each t
v(t,w) is !t-measurable
> O.
the Ito integral
6t V(s,w)dBs(oo) as a limit in L2(px) of sums ~ ei(w)~B.(w), where 1
1
- Bt . • w -+ ei( w) is !t.-measurable, consfruction is based on the Ito isometry:
ti
(1.19)
EX[]- v 2 (s,w)ds]
EX[(] v(x,w)dB s (w»2]
o
o
and the formula t
EX[(f v(s,w)dBs(w)] = 0
o
(1.10)
for all t.
Similarly one can define Ito integrals in higher dimensions: t fv(s,w)dB s
o
v
t m.nxm
when Bt is m-dimensional Brownian motion. A stochastic integral is a stochastic process form
Xt = Xo + where Xt , XO, u t
-b
t bt u(s,w)ds + 6 v(s,w)dB s
t m.n , v
c
Xt
on
the
(1.11)
m.nxm and u(s,w) is.1Jx 1-measurable,
u(s,w) ds < 00 a.s. An abbreviated version of (1.11) is (1. 12)
STOCHASTIC PROCESSES, INFINITESIMAL GENERA TORS AND FUNCTION THEORY
The Ito formula states that if in ]Rn and g t. CZ(]Rn, ]Rm) then Yt integral, given by
Xt =
143
is a stochastic integral g(X t ) is again a stochastic
(1. 13)
y(k) . h were t 1S coor d'1nate k
by the formulas dt'dB
(i) = t
. dB(i) = dt. i, and dB(i) t t We now turn to the Levy theorem. This result was first stated by P. Levy in 1948 [181 (in the complex plane), but - to the best of my knowledge - the first complete proof appeared in McKean (1969) [191. About 10 years later Bernard, Campbell and Davie [ 11 extended the result to higher dimensions. Both these proofs are based on stochastic integrals. j
,J.
T
1.2. Theorem (The McKean-Levy theorem).
Bernard-Campbell-Davie
extension
of
the
Let U C ]Rn be open and cp = (CP 1,;. ':" CPm) : U +]Rm be a C2-function. Let (Bt,~,PX) and (B t , ~,py) be Brownian motions in ]Rn and ]Rm, respectively. Then the following and (II» are equivalent:
«1)
(1)
cP (B t ) is - up to the exit time Brownian motion in ]Rm, except for More precisely, if we define
TU from U - again change in time scale.
=
T
a
(1.14)
then at is strictly increasing for a.a. wand cp*(w)
lim cp( Bt) ex i s t sa. e. on {w; a ( T)
< oo}
(1.15 )
t1T
And the process Mt(w,~),
(w,~) E~X~, defined by t
<
a (T ) (1. 16)
With probability law pX X pO coincides in law with (Le. has the same finite-dimensional distributions as) Brownian motion in !RID starting at CP(x).
144
(II)
B. 0KSENDAL
For all x, E: U we have (i)
(ii) (iii)
iV<11(x)i = iV¢j(x)i for all i,j, V¢i .~j = 0 if i f j (where· denotes the usual inner product in mn) [:, ¢i = 0 for all i
(1.17)
(where [:, denotes the
Laplacian Following Bernard, Campbell and Davie we will call functions ¢ satisfying (I) above Brownian path-preserving (BPP). Using (II) we now note the following: (i)
If n = m = 2 (and we identify m2 with () the only BPP functions are the analytic functions and the conjugate analytic functions.
(ii)
If n 2k, m 2 then every analytic function ¢ : U C ~k + ¢ is BPP. (But there many be others, see [1 J)
(iii)
If n = 2k, m = 2j > 2 then it is not the case that every analytic function ¢ : U C f;"K"+ cj is BPP.
During the last 10 years there have been many applications of this resul t to analytic functions, see for example Davis [ 9 J and Burkholder [2 J. Here we will try to illustrate the result by applying it to study the class H2(U) of analytic functions ¢: U + ¢ for a general open set U C a;n. (If U is the open unit disc in ¢, then H2(U) coincide with the classical H2 space). Before we define H2(U) let us recall that a stochastic interpretation of the Green function GU(x,y) = G(x,y) is that G(x,y)dy (where dy denotes Lebesgue measure) is the expected total length of time Brownian motion stays in dy before exiting from U, when starting at x. In other words, 0.18) for all f E: Co(U) (Le. f is continuous with compact support in U). Now let ¢ = (¢1, ¢2) : U C ICn + cr. be analytic and let W be open, we u, T TW. Then by Ito's formula ¢i (B T )
T
=
¢i (x) + f'i7¢i (Bs)dB s
i
= 1, 2
(1.19)
o
So by the Ito isometry t replaced by T)
(1.9)
and
(1.10)· (which also hold with
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
145
0.20 )
Therefore we have: 1.3.
Proposition. (i)
The following are equivalent:
2 ~up EX[I
<
00
for all x (; U
EX[a,u 1
<
00
for all x f
w-u
(ii )
(iii) (iv)
2
J I V
<
U
for all x (: U,
i
1,2
1
Proof. The equivalence of (i), (ii) and (iii) follows from 0.20) and 0.14), and it is easy to see that (i) is equivalent to (iv) by using the mean value property (1.6) for harmonic functions. 1.4. Definition. Let U C a;n be open and
i
J I Wi (y ) 2 d y
<
for i
0.21)
1, 2
U
Since
~
where ,
TU,
we
have
0.22) Therefore, a sufficient condition that an analytic function ¢ : U -+ a; belongs to H2(U) is that for y (:
a: .
See Burkholder [ 21 for an extension of this connection to HP; 0 < p < 00, in the case where U is the unit disc in 0: •
0.23 )
B. ~KSENDAL
146
a)
Boundary values and aSymptotic values
1.5. Definition. We say that
w = lim
The L~vy theorem immediately asymptotic values if
Theorem.
=
Let
lim t h(w)
E H2(U).
Since
EX[a
,]
existence
of
exists a.s. px, for all x E U.
<
The second assertion follows (l.S) of harmonic measure.
the
Then, with, = 'u,
In particular,
gives
00
we
values
have a
at
,
<
a.e.
00
point
a. s.
on
So by
aU
the
from the stochastic interpretation
Remarks. 1) This resul t holds more generally for finely harmonic functions satisfying (iii) in Proposition 1.3. See [21]. 2) The function
=
pX.[at
>0
o
It is well known that i f G c
for all x R2
(1. 24)
then G is polar iff cap(G) capacity, and if Gc mn with = 0 where Cn is the Newtonian n-capacity, i.e. the capacity associated to the kernel \x-y\ 2-n. If K C au is compact we let Aq, (K) denote the set of all asymptotic values of q, at K. The following theorem may be regarded as a partial extension of classical results in the unit disc due to Frostman, Nevanlinna and Tsuji (see Tsuji [25], p. 339).
= 0 where cap denotes logarithmic n 2. 3 then G is polar iff Cn (G)
147
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
1. 7. Theorem. Let Uc «;n be open, Kc au compact and !P. ~ H2 (U). Suppose A = A¢(K) has harmonic measure 0 wrt. ¢(U) \ A ~.~ (This occurs in particular if A is a polar set). Then K has harmonic measure 0 wrt. U. Proof. Let -I denote the first Then by hypothesis ~ = T¢(U) and
o using the notation we deduce that PX[:B:t ~ crT <
from
the
exit
theorem.
with ¢(B -1) (:A] = 0 crt
=
Hence px [ :as ~ T, crT <
"" ,
¢(Bs) (: A]
from
¢(U) \
A.
for all y (: ¢(U) \ A,
L~vy
where ¢(B T) is interpreted as ¢*
time
Since crT
<
T<jJ(U)
for all x, (: U,
lim ¢(B t ). t 1T
=0
and so < ""]
PX[BT (: K, crT
=0
for i f B (: K, crT < "" then T we have cr <""a.s. and hence T PX[B b)
T
E K]
=
0
¢(B
T
)
(: A a.s.
as asserted.
Removable singularities.
As an application we now prove that the polar sets are removable singularities for the space H2(U). For n = 1 this was first proved by Parreau [24] for all HP(U), p > O. The case n > 1 has been obtained by Jarvi [16] and also by Fuglede [14]. One can also extend the H2(U) result to the more general Space of finely harmonic functions satisfying (iii) in Proposition 1.3 (see [21]), but that requires more work. In the case where U is an open subset of «; Conway and Dudziak [ 6] have proved that the polar sets are exactly the removable singularity sets for the spaces HP(U), 0 < p < "", if one requires analytic extension with the same HP-norm.
1. 8. Theorem. Let U (: en be open, K a polar relatively closed subset of U and ¢ E H2 (U \ K). Then ¢ extends to an analytic function ~ in U (and ~ EH2(U». Proof. Let {Wk} be an increasing sequence of open sets with union U \ K. Let Tk = TWk . By Theorem 1.6 we know that ,1/Ik = ¢(B Lk ) -+ ¢* = lim ¢(B t ) a. s. since K is polar. We claim that in fact ttL
B. ¢>KSENDAL
148
in
for all x t
To see this we apply the Ito obtain that if k < m, i = 1, 2 EX [ (,j, . (B '+'1.
=
Tm
) -
,j,. (B '+'1.
Tk
formula
U\ and
0.25 )
K. the Ito
isometry
to
»2] =
Tm 2 EX[ J IV¢i (Bs) Ids] Tk
as m, k -..
co
since EX[o ] < co • Since T ¢ is harmonic in U \ K we have, using 0.25) for all x t U \ By the Harnack Therefore
inequalities
'1/
t
L2(px)
K
for all x E K as well.
(1.27)
x t U,
constitute a harmonic (and hence analytic) U. Note that since K is polar we have 2 I ¢liH2 (u\f<)
(1. 26)
extension
of ¢
to
2 TU\X 2 ifl EX[ JO I v¢i (Bs) Ids]
II ¢11!2 (U) . For results about removable singularities for the space of analytic functions with bounded Dirichlet integral see ¢ksendal & Stray [23]. In 1924 Rad6 proved that i f D = {I z I < l} c a; then a continuous function on D is analytic on D if it is analytic on D \ ¢-1(0). Later several extensions have been found (see Fuglede [14] and the list of references there). The version we present is based on the method used in 0ksendal [20]. 1.9.
Theorem.
(Extended Rad6 theorem)
Let IB be the open unit ball in a;n, K a relatively closed subset of IB and let ¢ tH2(IB \K). Put A = A¢(K). Suppose ¢CIB \K) \A '# 0 and A has harmonic measure 0 wrt. ¢(B\' K) \. A. Then ¢ extends to an analytic function on IB . Proof. Choose a component W of IB \K such that ¢ (W) \ A = 0. By the preceeding result K has harmonic measure o wrt. W. So Brownian motion starting from x E: W a. s. does not hit K before it hits 3IB. Hence W = IB \ K and K must be polar. But then K is a removable singularity set for ¢ by Theorem 1.8.
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
§2.
149
AN OPERATOR-CHARACTERIZATION OF PATH-PRESERVING FUNCTIONS
A natural question is: What is the reason for this connection between analytic functions and Brownian motion? We will give an answer to this question by turning to a more general situation: Suppose Xt is a Markov process on an open set U~ lRn , Yt a Markov process on V~ lRm and cp: U + V continuous. When will <j>(X t ) - up to the exit time from U for Xt - coincide in law with Yt, after a change of time scale? In order to explain this question (and the answer) precisely we need some definitions and notation: (We refer, for example, to Dynkin [12] for further details).
2.1. Definition. A Markov process on Uc lRn is a stochastic process Xt defined on (a family of) probability spaces (n,J,QX)xEU with state space U (Le. for each t w + Xt(w) is a map from n into U) together with an increasing family of a-algebras At~ J(t~O) such that (i)
Xt is At-measurable, for each t x + EX[f(X t ] (= Af(Xt(w»dQX) is Borel measurable for each boundeJ'Borel function f on U.
(iii)
(The Markov property) we nave ~a.s. fX)
For each t,h ~ 0, f bounded measurable
(2.1)
The process Xt is called (path- )continuous i f the paths t + Xt (w) are continuous a.s. QX for all x. A random time T : n + [0, ""] is called a (strict) stopping time (wrt. At) i f {W ; T( w)
< t}
E
.4
for all t > O.
(2.2)
'or example, if T = TU is the first exit time from an open set U C G for Xt , then T is a stopping time wrt. iJt . If T is a stopping time wrt. At, then itT is defined to be the a-algebra consisting of those N E iJ such that N
n
h ~ tl E
4
for all t.
14:.
(2.3)
is the a-algebra generated by {X s s < d then ./(T can be described simply as the a-algebra generated by {Xsi\T; s ~ O} . So reflects the history of the process Xt up to time T . The strong Markov property states that (iii) above holds with t replaced by any stopping time 't":
,If
k.
(iv)
EX[f(XT+h)
I
itT] = EXT[f(Xh)] a.s. QX, for all x.
B. ~KSENDAL
150
Processes satisfying (iv) are called strong Markov processes. To each Markov process we associate its infinitesimal operator or generator A defined by (Af)(x)
=
lim tlO
EX[f(X t )] - f(x) t if the limit exists.
(2.4)
Let DA (x) denote the set of functions f such that the limit (2.2) exists at x, DA the set of functions f which belongs to DA (x) for all x. Example 2.2. A way to construct Markov processes is by solving stochastic differential equations: Let b(x) E lRn, a(x) E lRnx m be continuous functions on lRn. An Ito diffusion is an .rt-adapted stochastic process Xt = X~ satisfying the stochastic differential equation
Xo = x,
(2.5)
(For where 'it is the a-algebra generated by {Bs s i t }. example, if b, a are bounded and Lipschitz-continuous, then a (unique) solution Xt of (2.5) can be constructed using a type of Picard iteration). Equipped with the measures QX defined by QX [X t 1 E G1, ... , Xtk E Gk]
= pO [Xx. E Gl' ... , XX tl
tk
E Gk] , (2. 6)
where pO is the probability law of Bt starting at 0, and the a-algebras At = 'it one can prove that Xt is a continuous strong Markov process with generator (Af)(x) where aij = ~(aa T)ij b = (bi)' Conversely, starting with a semi-elliptic 2nd order partial differential operator A as in (2.7) one can find a such that ~aa T = a = (aij) and hence construct via (2.5) an Ito diffusion whose generator is A, provided b and a satisfy the required conditions. See [22] for details. If Xt is an Ito diffusion with generator A then one can prove, using the Ito formula (1.13) and formula (2.7), that
for all 'Jt-stopping times T. This is the Dynkin formula. For simplicity we consider only Ito diffusions on. The time changes we will consider are of the form T
at
= ! ;>..(Xs)ds,
o
from
noW
151
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
where ,,(x) ~ 0 is a continuous function and the set N = {x ; ,,(x) = 0 } has empty X-fine interior, i.e. 'N = inf {t > 0 ; Xt iN}= 0 a.s. QX for all x. Such a A will be called a time change rate (for Xt ). Note that for such A the function t -+ Ot(w) is strictly increasing -1 for a.a. w. Its inverse will be denoted by 0 (w). t 2.3. Definition. Let (Xt'~ ,QX), (Yt,~ ,QY) be Ito diffusions on open sets U c mn, V c mm and let ¢: U -+ V be continuous. Then we say that ¢ is Xt - Yt path-preserving if there exists a time change 0t such that if we define, for W open, 'IV C U and, = 'w
I
¢ (X -1 ) °t Y
t-o
,
, >,
<
t t
0
(2.9)
0
with probability law QY given by, for ~ ~¢(U), EY[f1( Zt 1) ... EX[fl(¢(X -1» o t. (2.10)
... fk(Ytk-S)]s=o then (Zt ,QY) coincide in law (2.10) such that ¢(x) = y. in the matrix Gin (2.5) or we may assume that both QX and for suitable k.) We can now formulate path-preserving functions:
,
with (Yt,QY), for any choice of x in (Note that by adding some zero columns in the corresponding matrix for Yt , QY are measures on rl = C ( [0, (0), IRk)
a
the
operator
2.4. Theorem. (Csink and 0ksenda1 7 ] ). diffusions on open sets U c mn, V c mm with respectively.
characterization
Let Xt , Yt generators
be A,
of
Ito A
Then the following are equivalent: (i) (ii)
¢ is Xt -
Yt path-preserving
There exists a time change rate" such that for all f
E
C~(V)
(2.11)
Proof. (i) - (ii): Assume that ¢ is Xt - Yt path-preserving. By continuity it suffices to prove (2.11) for each x such that A(X) > O. If we let It denote the time changed process (2.12)
B. ~KSENDAL
152
then Xt is also an Ito diffusion infinitesimal operator of Xt we have
1 Ag(x) = \(x)
and
if
A
denotes
the
(2. 13)
A g(x)
(See Dynkin 1[12], p. 329).
So it suffices to prove that
A[f o¢](x) = A[f](¢(x»
(2.14)
,. where A is the infinitesimal operator of Yt . For convenience we drop the bar and simply ~rite Xt instead of Xt etc. Let y = ¢(x), x EWe W e U, W open, and T = TW. Then T is an It-stopping time. By Dynkin's formula (2.8) we have for f
f
C~(V) f(y)
tAT
A
+
6
EY[
(Af)(Ys)ds] = tAT f(y) + EO[ f (Af)(yy)ds],
°
where EO denotes the expectation wrt. pO. EY[f(ZtAT)] = EX[(f
°
s
Similarly,
¢)(X tAT )] = f(¢(x» tAT
EO[ f
°
(2. 15)
+ (2. 16)
(A[f
°
¢] )(Xx)ds]
s
The left hand sides of (2.15) and (2.16) coincide since ¢ Xt - Y t path-preserving and therefore by (2.l6)and (2.15)
A[f
° ¢]
(x)
lim iY[f(YtAJ] - f(y) tW
y=¢(x).
EO[t/\T]
(ii ~ (i): Assume that (ii) holds.
Let Bt
-1 t and Xt
-0
Then by the argument above fCp(x»
BtAo + EO[ J TA[f o¢l~~)dsl
o
BtAGr = f( ¢(x»+ EO[ f A:(Xx)(Af)(¢(~»dsl s s d -1 1 .... at is strictly increasing we have dt (0 t ) = \(~) s
o
Since t
>
°
is
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
for a.a. t (a.s. pO) so by the substitution s =
o
EY[f(Ztl\a )1 = f(¢(x» ,
+ E [
t/'o,
f 0
A
0-
u
153
1 we obtain
-X
(Af)(¢(Xs»dsl = (2.1])
~ f(¢(x)) + EY[] (Af)(Z )ds
o
s
Similarly, Yt we get:
from
(2.10)
EY [ f (Z t)
X
.
and Dynkin's
formula
(2.8)
applied
to
1
,
{t>o }
(2.18)
Now EY[f(Ztl\ojl
= EY[f(Zt) . X{t~o,}l + EY[(f(Z,) . X{t>o }l ,
(2.19)
EY[f(Zt) . X{t
,
,
X{t>o } 1
Therefore, by adding (2.17) and (2.18) we get EY[f(Zt)l + EX [f(¢(X,» t
_
. x{t>o}l = f(¢(x»
,
+
A
6EY[(Af)(ZS)(X{t
"
} + X{t>o })l ds + EX[f(¢(X,»
. x{ t>o }l
,
or
EY[f(Zt)l
=
+ JE1Af)(Zs)dS
f(¢(x»
(2.20)
o
Similarly, Dynkin's formula applied to Yt gives directly A
t
A
EY[f(Yt)l = f(y) + fE~Af)(Ys)ds
(2.21)
o
So the two families of bounded linear functionals u t0) : CO(V) t ~ 0, i = 1, 2, defined by f
both satisfy the equation
E CO( v)
-+ JR,
154
B. (>KSENDAL
f
E CO(V)
(2.22)
So by uniqueness (see the next lenuna) u(l)(f) = u (2) (f) for all t t i. e. EY [f(Zt)] = EY[f(Y t )] for all
f E CO(V) f
E CO(V).
Similarly one proves by induction that
-
-
EY[fl(Ztl) ... fk(Ztk.:tt)] = EY[fl(Y tl ) ... fk(Ytk.t t )]
i
for all 0 tl < t2 < ... tk.,. t ~ 0 and fl, ... , fk E CO(V). That completes the proof that (Zt, QY) is identical in law with (Y t , QY), except for the uniqueness claimed above: This next lemma works in general if A is the infinitesimal generator of a contraction semi group {T t } of operators on Cb(V), the bounded continuous functions on V, equipped with the uniform topology. (See Lamperti [17], Ch. Then fJA denotes the (dense) set
n.
of functions f E Cb(V) such that Aft= {[Ttf - f] converges uniformlyon V as t -+ O. In our case the contraction semigroup is given by T t f = E'[f(Y t )], and it can be seen using the Dynkin formula (2.8) and the Lipscitz condition on band 0 in (2.5) that C2 c 2A
o
A'
2.5 Lenuna.
(Uniqueness Lenuna)
Let {w t } be a family Cb(G) intoN satisfying
of
bounded
for all b E
linear
functionals
from
.1);\
and for some constants C, m Wt f -+ 0
(ii)
for all
Then Wt g = 0 Proof: nt
Choose a
e- at
t -+ 0+
as
>m
for all f
E;
<
0
2JA •
g E Cb (G ) , t > O.
and put
W't
-
A)-l (a I Let Ra Lamperti [171 , p. 142) . Then
Cb (V) -+ 5J A be the resolvent of A Choose g E Cb (V) and put f = Ra g E
(see
.1t.
ISS
STOCHASTIC PROCESSES,INFINITESIMAL GENERATORS AND FUNCTION THEORY
~(nt
f)
=
+ e- at Wt Af = nt(-aI + A)f
(-a)nt f
n t R-1 a f
n t R-1 a Ra g = - nt g.
Hence t
=-
f ns g ds
o
By
(0 nt f
+ 0
using (ii)
nt f
as
t +
co
and we conclude that for all a> m
This implies that Ws g
=0
Remark:
for all s
~
0, as claimed.
Note that it is a consequence of (2.11) that
Ah :: 0 in G (open in V)
~ A[ho» ::
0 in >-l(G), h f
C~(V)
(2.23)
In fact, it turns out that (2.23) is actually equivalent to (2.11), under mild conditions on > on Yt (see [ 7 ) . If we call the functions h such that Ah:: 0 in G for A-harmonic functions and similarly with A, then (2.23) says that > is an X-X harmonic morphism. See Constantinescu & Cornea [4) or Fug1ede [13) for more information about such functions. For a stochastic interpretation in a more general setting see Csink & 0ksenda1 [8). It turns out that i f > 'satisfies (ii) of Therem 2.4 then > automatically has boundary values on aU, in the following sense: Suppose >
2.6 Theorem. (Cs ink & 0ksendal) [7) . satisfies (ii) of Theorem 2.4. Then >*
lim > (X t )
exists a.s. on{a,U
<
co
f
C2(U,V)
}
U
t
where t
f A(X S )ds.
at
o
Moreover, if we define Zt as in (2.9), (2.10) but with, = and >(X ) replaced by >* then Zt is identical in law to Yt .
'u
l'
Proof. The basic ingredient in the proof is the Prohorov theorem: Equip sl = c ([ 0, co), lRm) with the topology 'of uniform convergence on b~und~d interva1s.~ !hen for all e: > 0 there exists a compact KC n such that PY(K) > 1 - e:. We refer to [ 7) for details.
B. ~KSENDAL
156
We end this section by showing that the Levy theorem as a special case: Choose Xt = Bt , Yt = Bt to be Brownian mm, respectively. Then the generators Laplacian operators ~~ inm n , mm and (2.11)
Theorem 2.4 contains motions in m n , of Xt , Yy are becomes
the
By comparing terms on both sides we easily see that this is equivalent to t~e conditions (1. 17) in the Levy theorem, with >.. (x) = IV¢i (x) I ; i = 1, ... , n. Finally, since the critical points of a (classical) harmonic morphism constitute a polar set (Fuglede [13), p. 116) the function>.. is a time change rate, and the Levy theorem follows from Theorems 2.4 and 2.6.
§3.
APPLICATIONS
Based on the operator characterization of path-preserving functions it is natural to adopt the following point of view: Given a function ¢ whose properties we want to investigate, we try to find semi-elliptic 2nd order partial differential operators A, A with continuous coefficients such that for f
E C2
o A,
for some continuous function >..(x) < O. If A, are "reasonable" we then conclude tha~ there exist Markov processes Xt , Yt (whose generators are A, A respectively) such that ¢ is Xt Yt path-preserving (with time change rate >..). Based on known behaviour of the paths of Xt , Yt we may then deduce properties about ¢, like one has done for analytic functions ¢: Uc a;n -;- a; in the case when Xt and Yt are Brownian motions. Here are 3 reasons why it is worthwhile to consider the general path-preserving theorem 2.4 rather than just the Levy theorem: I f ¢: uc: a;n -;- Ck with k > 1 is analytic, then ¢ is not necessarily BPP, not even if ¢ is biholomorphic. Thus the Levy theorem is not the right tool for studying these functions.
2)
Even in the case when the function ¢ is BPP, there may be other processes Xt , Yt as well such that ¢ is Xt - Yt path-preserving, and these processes may perhaps be chosen to serve our purpose better than Brownian motion. In a) below we give an example by considering the conditional Brownian motion. (However, if ¢ is analytic then such processes Xt , Yt must depend on ¢ , unless they are just time changed Brownian motions. See [ 7]).
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
3)
a)
157
It is of interest to have an approach which is not limited to analytic functions. See b) below.
Analytic functions and conditional Brownian motion
Conditional Brownian motion was introduced by Ooob [101 and a detailed account of this process can be found in his latest book [ 111 • If h > 0 is a harmonic func tion on a domain U C lRn, then the h-conditional Browning motion, denoted by {B~\>O' is a Markov process in U (possibly with a whose generator is given by (I,
h f
= ;ti hf 1 =
(I,f
+
h
One way to construct
B~
lire
finite
time l;.)
Vb' 'Vi h
(3.1)
is as follows:
Let
Uj
be an increasing
<»
sequence of domains such that UJ·
C
U,
.U
U
J=l
UJ' •
For
each
j let a(j) E CO(U, lRhxn) and b(j) E CoW" Rn) satisfy Jj)(x)
=
(the n Xn identity matrix) and b (j ) (x)
Then
= \}:(~x/
if
x (: Uj .
In
let X(j) be the solution of the stochastic differential equation t dX(j)
(3.2)
t
Let Tj = inf {t > 0; x~)f/. Uj} and put B~ for t < l; by setting Bh
t
= Xj
t
if
t
l;
lim Tj. j-+-
< Tj
(3.3)
Xj +1 (This makes sense since Xj if t < Tj) . t t 2 C function g h-harmonic in an open subset W of U i f (l,h g == 0
i. e.
Then define
co
l\[hg1
==
0
We
call
a
in W.
By Dynkin's formula we see that (3.4)
for all bounded h-harmonic g E C2 and all B~-stoPPing times T In particular, if we apply this to g = h we see that for all
E
> O.
<
l;.
(3.5)
B. ~KSENDAL
158
Therefore, if h extends continuously to an open subset L of au and is 0 there we a.s. do not have lim Bh E L. In particular, -ttr t i f h extends continuously to au \ { y} and is 0 there (and then necessarily unbounded at y) then h
B t -+- y
a.s.
as ttl;;. h
Remark. For any compact KCU Bt only spends a finite time in K, a.s. This h is because the h-Green function (i. e. the Green function for Bt ) has the form
b EX[XK(B ht) co
Gh(x,K) =
.
-~ h(x)
X{t
t0
(3.7)
EX[h(Bt)XK(B t ) . X{ t<,.. }ldt .,
which is finite provided U is Greenian" i.e. has a finite Green function for Bt . For n > 3 all domains ucmn are Greenian, while for n = 2 uc m 2 is Greenian iff U has a non-polar complement. We will throughout a) assume that U is Greenian. Intuitively Bh can be viewed as a Brownian motion driven to exit from U ~t the points where h is nonzero. (This is taken care of by the first order "drift" term of lIh in (3.1». Our next result is that the Brownian path-preserving functions also are "conditional path-preserving", in the following sense: 3.1. Theorem. Let U be a domain inmn and let cp : U -+- Vc. mm be a BPP function. Let h > 0 be harmonic on V and put k = h 0 cpo Then cp is B~ - B~ path-preserving, with time change rate A = I~CPll 2 as in the L~vy theorem. Proof. This established that
follows
from
Theorem
2.4
once
we
have
(3.8) And (3.8) follows directly from (3.1) and (2.24):
Using Theorem 2.6 we now conclude: 3.2.
Corollary.
Let cp, h, k be as in Theorem 3.1.
Put t ~ 1;;'
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
159
Then exists a.s. on {a
k I;;
<
co}.
As an illustration we use this to prove the following result, which gives a condition for when a cluster value of a BPP function is an asymptotic value. Let us call a point v E av strongly accessible for V is there exists a harmonic function h > 0 in V such that h(x) -+ 0 as x -7 y, for all y E av \ {v}. And we say that w is an extended asymptotic value for a functioncpdefined on U if there exists a curve y(t) : [0, (0) -+ U such that lim
cp(y(t»
= w.
t-+ oc
3.3. Corollary. Let UC mn be a Greenian domain and let cp : U -+ V = c!>(U) c: mm be BPP and proper (Le. cp-l(K) is compact for all compact K C V). Suppose w is a cluster value of cp and that w is a strongly accessible point for V. Then w is an extended asymptotic value for cpo Proof. Since w is strongly accessbile for V we can choose a harmonic f~.mction h > 0 in V such that h -+ 0 on av, \ { w} . Put k = h 0 cpo Then by Corollary 3.2 we have that cp* = lim
CP(B~)
exists a.s. on
ttl;;
{l
<
co}.
I;;
SincR cp is proper we must havR cp* E aV an~ hence cp* '" w a.s. on {a < oc} by Theorem 3.1. If a = co then cp(B k -1) ; 0 i t <'"' , I;;
I;;
(at)
must constitute the complete path of Bh. Since a.s. end up at the point w we concludetthat
CP(B~) = w
lim
these
paths
a.s.
ttl;;
b) Finally we turn to the question: What continuous functions are Xt Yt path-preserving, for suitable choices of Markov processes Xt , Yt ? In view of the operator characterization (Theorem 24) this question is closely related to the problem of finding (semi-elliptic 2nd order partial differential) operators A,A such that A[f
0
cp] = A[f]
0
cp
for all f E
Co2 .
Therefore a partial answer is given by the following which is due to A. M. Davie (private communication):
resul t,
B. 0KSENDAL
160
3.4. Theorem. Let Uc: ]Rn be open and cp E Ck(U,]Rm) for k = max(2, n - min(n,m) + 1). Assume that cp is proper, i.e. cp-l(K) is compact for all compact KCcp(U). Then there exist non-zero semielliptic 2nd order partial differential operators A, on U, ]Rm such that
A
A[f
cp] = A[f]
0
2 for all f E Co
cp
0
0.11)
Proof. Let p be the maximum rank of the derivative matrix cp' on U and put rank ( cp ')
}
Then E is closed. For any x E U \ E we can make a C2 change of coordinates near x and cp(x) such that in a neighborhood of x cp has the form 0.11) Since cp is proper, i t follows from 0.11) that CP(U) \ CP(E) is locally a finite union of p-dimensional submanifolds of ]Rm. Let W be the set of points y E cp(U) '. cp(E) such that y has a neighborhood N such that N " ( cp (U) \ cp(E» is a p-dimensional manifold. Then W is a p-dimensional submanifold of]Rm and W is relatively open and dense in cp (U) \ cp(E). (Intuitively, W is <jl(U) \ CP(E) with the "crossings" removed). Choose y, E W. Since cp-l(y) is compact, we can cover<jl-l(y) by open sets Gl, ... , Gj for each of which 0.11) is~ valid. Then H = "CP(Gi) is a neighborhood of y in W. Let A be an elliptic operator on W with support on a compact subset of From (3.11) we can find for each Gi a semielliptic operator Ai on Gi such that
H.
Ad f
0
cp]
=
A[f]
cp
0
for all f, (::
C~ (H)
Let 1jJl, ... 1jJs be smooth functions with supp l:1jJi = 1
cp-l (supp
on
~'ic;.
O. 12) Gi and
A).
Then 0.13 ) is a well-defined semielliptic operator and A[f
0
<jl] = A[f]
0
<jl
To obtain an operator A which is non-trivial everywhere in W one can take a sequence A(r) of operators as above whose support cover W with corresponding A(r) and put
STOCHASTIC PROCESSES, INFINITESIMAL GENERATORS AND FUNCTION THEORY
161
where u r + 0 very fast. This construction would break down if ~(U) \ ~(E) = 0. But since ~ E Ck with k 2. n - p + 1 we get by Sard I s theorem that ~(E) has p-dimensional Hausdorff measure O. Therefore we in fact have A non-zero on a dense subset of U \ E. Acknowledgements. I wish to thank the London Mathematical Society and the NATO Advanced Study Institute for their financial support. And I am grateful to J. J. Conway, A. M. Davie and J. L. Doob for valuable conversations during the preparation of these notes. REFERENCES [1]
A. Bernard, E. A. Campbell and A. M. Davie: Brownian motion and generalized analytic and inner functions. Ann. Inst. Fourier 29 (1979), 207-228.
[2]
D. Burkholder: Exit times of Brownian motion. Advances in Math. 26 (1977), 182-205.
[3]
J. Bliedtner and W. Hansen:
[4]
C. Constantinescu and A. Cornea: Compactifications of harmonic spaces. Nagoya Math. J. 25 (1965), 1-57.
[5]
C. Constantinescu and A. Cornea: Potential Harmonic Spaces. Springer-Verlag 1972.
[6]
J. P. Conway and for HP functions.
[7]
L. Csink and B. ~ksendal: Stochastic harmonic morphisms: Functions mapping the paths of one diffusion into the paths of another. Ann. Inst. Fourier 33 (1983), 219-240.
[8]
L. Csink and B. ~ksenda1: A stochastic of harmonic spaces. (To appear).
characterization
[9]
B. Davis: Brownian motion and analytic Annals of Probability 7 (1979), 913-932.
functions.
Markov processes and harmonic spaces. Z. Wahrsheinlichkeitstheorie verw. Gebiete 42 (1978), 309-325.
J. J. Dudziak: Preprint 1984.
Removable
Theory
on
singularities
The
[10] J. L. Doob: Conditional Browning motion and the boundary limits of harmonic functions. Bull. Soc. Math. France 85 (1957), 431-458.
B.~KSENDAL
162
[11] J. L. Doob: Classical Potential Theory and Counterpart. Springer-Verlag 1984. [12] E. B. Dynkin:
Markov Processes I, II.
Its Probabilistic
Springer-Verlag 1965.
[13] B. Fuglede: Harmonic morphisms between Riemannian manifolds. Ann. Inst. Fourier 28 (1978), 107-144. [14] B. Fuglede: Value distribution of harmonic and finely harmonic morphisms and applications in complex analysis. Kobenhavns Universitet Matematisk Institut Preprint Series 10 (1984). [15] L. L. Helms:
Introduction to Potential Theory.
Wiley 1969.
[16] P. Jarvi: Removable singularities for HP functions. Amer. Math. Soc. 86 (1982), 596-598. [17] J. Lamperti:
Stochastic Processes.
Proc.
Springer-Verlag 1977.
[18] P. Levy: Processus Stochastiques et Mouvement Brownien. Gauthier-Villars 1948. [19] H. P. McKean:
Stochastic Integrals.
Academic Press 1969.
[20] B. ~ksendal: A stochastic proof of an extension of a theorem of Rado. Proc. Edinburgh Math. Soc. 26 (1983), 333-336. [21] B. ~ksendal: Finely harmonic functions with finite Dirichlet integral with respect to the Green measure. To appear in Trans. Amer. Math. Soc. [22] B. ~ksendal: An Introduction to Stochastic Differential Equations with Applications. (To appear on Springer-Verlag). [23] B. ~ksendal and A. Stray: Removable singularities for analytic functions with bounded Dirichlet integral. (To appear). [24] M. Parreau: Sur les moyennes des fonctions harmoniques et la classification des surfaces de Riemann. Ann. Inst. Fourier 3 (1951), 103-197. [25] M. Tsuji: Potential Theory in Modern Function Theory. Maruzen 1959.
PARACOMMUTATORS AND MINIMAL SPACES
Jack Peetre University of Lund, Box 118, S-22100 Lund, Sweden
About these lectures The purposes of these lectures is twofold. First of all we would llke to give a reasonably selfcontained exposition of some topics of Analysis which I happen to possess some first hand acquaintance with. Thus in Lecture 1 I review some of the sal lent facts about (mostly) real interpolation spaces, and similarly, in Lecture 2 I discuss briefly the theory of Besov spaces. Although all of this material is in principle available in book form I venture the belief that these ideas, in itself rather straightforward, have not yet received the attention among analysts, I think, they justly deserve, which therefore gives a motivation for taking them up for a quick discussion once more. In the rest of the space allocated to me i then turn my attention to some more recent developments in Analysis which I have had some contact with. Thus in Lecture 3 I discuss recent work of Arazy and Fisher on Mobius invariant spaces of holomorphic functions over the unit disk. In particular, the minimal space introduced by them lends itself to a very simple proof of the trace ideal criterion for Hankel operators, associated with the name of Peller. Now on the real line the theory of Hankel operators is essentially equivalent to the theory of commutators of CalderonZygmund operators studied by Janson and others. In Lecture 4 we discuss even wilder fellows termed paracommutators, by analogx with paramultiplication in the sense of Bony, which live on R • Finally, in the short Lecture 5 we make some remarks on multilinear forms over Hilbert spaces, a theory which is still in a rather embryonic state, motivated by the observation that paracommutators (and Hankel operators too) really should be viewed as forms, not operators. It is clear that in this second part of the lectures, owing to the more advanced character of the subject, we have to be even more sketchy referring to the literature for most proofs. Also some of the material has not been published before. Lectures 1 and 2 are also accompanied each by a set of notes, mainly giving historical (bibliographical) information. To facilitate reading (and for easier cross reference) we have divided up the lectures in Sections. A Section which contains some new (newer) material has got a star * in front of it. 163
164
J. PEETRE
LECTURE 1. INTERPOLATION SPACES.
l. The theory of 1nterpolation spaces 1S basically a theory of fam1lies of Banach spaces ("scales of spaces" in a more techn1cal sense, see [K3)). Thus attention is fixed on the common features of fam1lies of spaces such as LP , c a , •.. , with poss1ble applicat10ns to other branches of Analys1s in mind. Such famil1es have 1n common the remarkable property that they can be reconstructed from the "endpoint" spaces, by an abstract procedure known as "interpolation". Remarks. 1. Most of what we say here 1S also val1d 1n the more general case of quas1-Banach spaces. 2. Of course, not all famil1es ar1sing in "nature" come under this heading. As an example of a family of spaces which doesAnot possess the 1nterpolation property we mention the spaces L P of Stampacchia; th1S, of course, does not imply that such spaces are w1thout any interest.
l. The by far most 1mportant 1nterpolation method 1S undoubtedly the complex method (the spaces [AO,A )8)' a modern offshoot of the classical Riesz-Thor1n theorem. We sfiaJl not glve the deta1ls here but refer to, for instance, [B2), chap. 4. Instead we shall dwell at some length on 1tS compet1tor, the real method (the spaces (Ao,Al)e ), 1n a slm1lar way connected w1th the theorem of Marcinkiewic~. Most proofs om1tted here aga1n w1ll be easily found 1n [B2), chap.3. 3. Let there thus be glven two Banach spaces AO and Al both continuously 1mbedded 1n one and the same Hausdorff topological vector space; techn1cally speaking, we thus have a compatible pair of Banach spaces, notation: A = (AO,A l ). Tgen one can do the follow1ng construct1on: 1 For any 0 < t < 00 and a AO + Al (hul 1 of AO and Al in A; a E. AO + Al <=>] a i e. A1 (1 = 0,1) such that a = a O + all we set ("K functional") K(t,a)
K(t,a;A)
inf a=aO+a l aO€'AO' alE: Al
2 0 For any 0 < t < 00 and a E. AO AI) we set ("J functional") J(t,a) = J(t,a;A)
n
Al (lntersection of AO and
maxlilallA,lIallA
o
) 1
PARACOMMUTATORS AND MINIMAL SPACES
165
Thus we have in each of the spaces AO + A and A n Al a one parameter family of equivalent norms. 1 Motivation for forming K(t,a), say: K<1,a) is the natural norm in A + AI. So K(t,a) is what is obtained when one replaces the given normOin Al by a multiple.
°
~
It is easy to prove that
(1)
K(t,a)
~
max(l,t/s)K(s,a),
(2 )
J(t, a)
~
max(l, t/s)J(s, a),
(3)
K(t,a)
~
min(l,t/s)J(s,a);
in particular, inequalities (1) and (2) make precise the above statement about norm equivalence. The K an~ J functionals are related by duality. More precisely, assume that A = (A O,A 1) is a regular pair, that is, the intersection AO n Al is dense in both AO a~d AI' so that _tfe dual pair i* = (A~,A;) can be formed. Then e.g. K(t,a;A) ani::l J<1/t,a;A ) are dual norms, on AO + Al
. and AQ* n Al* (~ (A O + AI) * ) respectively. Indeed, directly from the definitions if follows that I 1
-* )K(t, a),
J(l/t, a';A
~
which gives half of the statement. The converse is equally simple: Every continuous linear functional on AO + Al arises from an element of
* --* A* O n AI. One can likewise prove that J(t,a;A) and K<1/t,a';A ) are dual norms But this is (although) simple less elementary, since the converse now requires the Hahn-Banach theorem. 5. With the aid of the K and J functional one can now define interpolation spaces (K and J spaces). If < e < 1, 1 ~ q ~ 00 we define a "function norm"
°
f
00
(t-ecp(t»q dt/t)l/q,
o
cp any positive measurable function on (0,00). Using the K functional we then obtain the space
and, similarly using the J functional, the space 3 u = u(t).
J.PEETRE
166
a measurable £unction on (0,.) with values in AO a
•
=J
o
n
Al ,
u(t) dt/t (convergence in AO + Al ) and
+eq(J(t,U(t»
< .}.
It is easy to see that these are Banach spaces again continously imbedded in the "containing" spaceJ\.
PARACOMMUTATORS AND MINIMAL SPACES
167
3. It is often technically advantageous (and, as we have seen, sometimes necessary) to use a "discrete" definition. In fact, one can put the definitions in the following form:
[~ (2 v9 K(2 v ,a»q]l/q v=-.
• v v=-. ,
3 u = (u)
< ••
a doubly infinite
sequence with values in AO n Al such that a =
~
v=-.
(convergence in AO + Al ) and
[~ (2 V9 J(2 v ,U v )q]l/q v=-·
u
v
< ••
Thus, formally speaking, the multiplicative group of positive reals.&... = (0,.) has been replaced by the additive group of integers b
K (t, f)
= K (t, f; L l, L.)
I• f* (x)
=
dx,
o
where f * is the decreasing rearrangement of f on <0,.). Indeed, for the proof it suffices to consider the casl Z = (0,.), ~ = the usual measure and f itself decreasing, so that f = f • Then the "optimal" decomposition f = fO + fJ. in the competition for the K functional is obtained by "slicing" f ftorisontally at a certain level:
fO(x) = f(x) - f(t) if x
~
t,
0 if x > t;
flex) = f(x) - fO(x).
0"
From this formula it follows that (L I ,L.) 9( = Lpq (Lorentz space) for lip = (1 - 9)/1 + 91. = 11 ; 9. (Proof Hardy's inequality.) In particular, we have thus (L ,L )sp = LP in the same hypothesis (lip = 1 - 9).
J. PEETRE
168
Again this formula can be generalized in many dl~ecUons. FOi instance, let H be a Hilbert space and consider the pair (S ,B) where S is the (normed) ideal of trace class 1 (nuclear) operators in Hand B the ring of all bounded operators. Then one has an analogous formula for the K cLuncUonals with the decreasing rearrangement of the sequence (sn (T» n=O of Schmidt (approximation) num~ers of a ~iven operator T. As a consequence one gets, for instance, (S ,B) = S~ where now sP is the ideal of trace (Schatten - von Neumann) cPPss p operators; the full result is about the space sp,q ("Lorentz-trace class"). Instead of Bone 00 . can also take S , the Ideal of all compact operators. 1 1 In another direction, we may replace the space L by H
a
1
K(t,f) = K(t,f;C ,C )
~
w(t,f)
where w(t,f) = sup I hi
0100
is the space COl of Holder continuous functions
(f E COI~w(t,f) = O(t 01». Since it is so simple and also the paradigm for many similar proofs let us indicate (at least the essentials of) the proof. One direction is easy: Since obviously
we get at once
w(t,f)
~
2K(t,f).
For the converse we have to produce an "approximately" optimal decomposition f = fO + fl' Our choice is t
fO(x)
= l/tJ (f(x)
a
- f(x + h) )dh,
f1 (x) = f(x)
(This is essential the Steklov decomposition.) It is clear that
- fO(X)'
169
PARACOMMUTATORS AND MINIMAL SPACES
But by partial integration f 1'(x) = (f(x + t) - f(x»/t. Therefore we also find
We conclude that
Notice that we have here essentiall y onl y used that C 1 is the domain oh the infinitesimal generator of a (one parameter) semi group acting in C , so the argument immediately extends to a more general abstract situation of a pair (E,D(A», where D(A) denotes the domain of the infinitesimal generator A of a semi group G(t) acting in a given Banach space E. Around 1960 Lions made a comprehensive study of interpolation of "semi group spaces", seminal for all later developments [L3]. Some of Lions's original work is reproduced in [A1]. And we should also not forget to Q...uote Grisvard's thesis [G41. 3. (LP,LP(w», a pair of weighted LP spaces (f E LP(w)<~wf E L P, w a positive "weight" function). To get an exact result we have to mo
a)
(and similarly J
p
(t,
a)
= J
p
(t, aiA)
+ tPllall~
)l/p).
1
Clearly the two families K(t,a) and K (t,a) are uniformly equivalent in an obvious sense. In the case at hand t!his gives
from ~hich again one ma~ ded~ce that (LP~LP(W»9 = L P (w 9 ). ThiS result (due to FOlalj-Llons [F41) IS conne8ted with the theorem of Stein-Weiss [571. The problem of interpolation of a pair (E,D(A», E an arbitrary Hilbert space, D(A) the domain of a self-adjoint operator A in E, can be reduced (via the spectral theorem) to the special case p = 2; this is the subject of Lions'S first paper on interpolation (1958), Which essentially establishes the fact that (E,D(A»1I2 2 =.D(A ). A related problem, which can be treated in esilenllall y the same manner (as interpolation of pweighted LP ), is the problem of interpolation of the pair (LP(AO),L (AI» (vector valued functions). One still has an explicit formula for the K uncUonal and as a consequence holds
[f-H
J. PEETRE
170
If we allow the p's to vary, one can still prove (by a quite different
route) that
provided lip = <1 - a)/pO + alp (a E (0,1». Note that in all these results q = p is adjusted to the si!uation at hand. The case of general q is much harder. (For the latest results on this question we refer to [PH].)
7. After this long digression (concrete examples of K functionals) we return to the general case. We would like to say a few words about the proof of the equivalence theorem (formulated in § 5). It is based on a technical result.
FUNDAMENTAL LEMMA (baby version). Let A = (AO,A I ) be any pair and let a be an element in the (norm) closure of AO n in Ao + At' Then there exists a measurable function with values in 0 n Al sucn that a = f~ u(t) dt/t (convergence in AO + AI) and such that
Al
(4)
J(t, u(t»
~
constant'K(t, a)
(t E
(0,00)
a. e. ).
Let us sketTch its proof for the corresponding discrete version (i.e. we take t !If 2 , k E l): Given a we would like to represent it in the form a = :E kEZ uk (convergence in AO + AI) with uk E AO n AI' This is
-
easy. For any k we pick (see the definition of the K functional> a Ok E AO and a lk E Al such that 2K (t, a).
laOk" A
o
Putting uk = a k - a O we get the desired decomposition, along with the estimate dke +Jiscre~e counterpart of (4» (4' )
k J (2 , uk)
~
k
4K (2 , a ).
#
The proof of the equivalence theorem is now easy. Indeed, the inclusion Aa. c: Aa .J is an immediate consequence of the fundamental lemma. Fo~'lhe con~"rse we have to involve inequality (3): From a = f~ u(t) dt/t we deduce that (5)
K(t,a)
~
00
J
min(l,t/s)J(s,u(s»
ds/s,
o
so if +aq(J(s,U(s» < 00 we readily conclude that +aq(K(t,a) < 00.#
171
PARACOMMUTATORS AND MINIMAL SPACES
Remark. The fundamental lemma, which has been along since the early 60's, has recently been strengthened by Brudnyr and Kruglyak [B8] who established the following remarkable result. FUNDAMENTAL LEMMA (strong form). Assume that A is "mutually closed" <1. e., each space Ai <1 = 0,1) is its own relative (Gagliardo) closure in AO + A 1). Then one has an analogous statement with (4) replaced by 00
(4') fmin( 1, t/s)J(s, u(s) )ds/s ~ const·K(t, a)
(t E (0,00».
o This result places a prominent rale in more recent developments in interpolation. In what follows we shall however not require it. 8. The equivalence theorem has a very important extesion known the reiteration theorem. We begin by a definition. Recall_that a Banach space A is said to an intermediate space for the pair A = (Alt A 1) if An n Al c: A c: Ao + algebraically and topologically (§ 5). We say tnat A is of class where 0 ~ 9 ~ 1 is a given number, if, on the one hand,
K(t,a)
~
9
Ct UaU A £or all a E A,
t
as be Al 9,
E (0,00)
and, on the other hand,
DaU A
~
Ct
-9
J(t,a) £or all a E AO n A1 ,
t
E (0,00).
The last thing can also be stated, eliminating the J functional, as
with the same qualifications. It is easy to verify that, if 0 < 9 < 1, A being of class e is equivalent to the two imbeddings A c: A c: A , so in this case this is indeed much more than to say t~ll\. A is jus~':n intermediate space. Exame,le. The complex interpolation space Ae = [AO,A 1 ] e is of class e. Also Aeq is of class 9, for any q. Now we can state the REITERATION THEOREM. Let the Banach space E and E be of class eO and e 1, respectively, where 0 < eO < e 1 < 1. ~en we1 have, up to equi valence of norm,
J. PEETRE
172
for a = (1 - A)a 0 + Aa l' 1 :S q :S
00.
For the proof se [Bl 1, chap. 3, p. 50-51. The following picture is useful to have in mind (each of our spaces is thought to have a "weight", a number in the interval [0,11):
o
1
In particular, the theorem is applicable in the following two cases (see the preceding example): 10 E. = [A O,A l 1a. . 1 :As an il ustration, we shall show how to improve upon a previous example (example 1 of § 6). Example. By that example we know that (L l ,LOII )a = L Pq if lip = 1 a, 0 < a < 1. It l follows that the space LPq (Lorentz ~ace) is of class aq for the pair (L ,LOll). Thus we may conclude that (L P Oq O,L P l q 1>a = LP if lip = (1 - a)/po + a/pl' 0 < a < 1, PO Pl. This is relateWto the theorem by Marcinldewicz.
*
9. An important complement to the reiteration theorem again has been obtained by T. Wolff [W 1 • Let A l , A2, A3, A4 be Banach space all four continuously imbedded in one and the same (farge) Hausdorff topological vector space (so that (A 1,A 4 ) forms a compatible pair etc.). Let further a, 1'), A, ~ be parameters subject to the following conditions: 0 < a < I') < 1, a = AI'), I') = (1 -
~)a
+
~.
WOLFF'S THEOREM. Assume that A2 = (A l ,A 3 )Ap and A3 Then A2 = (A l ,A 4 )ap and A3 = (A l ,A 4 )l')p· Assigning weights 0, a, the following picture:
o
a
1'),
= (A2,A4)~q
1 to A l , A2, A3, A4 respectively we have 1
Thus, intuitively speaking, Wolff's theorem allows us to patch together a "large" scale out of "small" scales, in an analogous manner as one in Differential Geometry, say, manifactures a whole manifold out of small coordinate patche~. The proof is, surprisingly, quite straightforward. In one direction:
we readily deduce that
PARACOMMUTATORS AND MINIMAL SPACES
173
whence (A 1,A 4 )e1 c A2 • The other direction: If a 2 E A2. with II aU A = 1 then by the definition of the K functional we can write (for t fixed) ~ = a + a with 213
Similarly, we can write (for any s) a 3 = a' 2 + a 4 with II a 4 UA
<_ C·s~-lt~-l.
4
We adjust s to t by requiring that s~· t~-1 l1li 112· previous C. We thr see that we have a 2 = a' 2 + b with II a' 2"A2 ~ 1/2, K(t,b;A l ,A4 ) ~ C· t • Now we repeat the same argument with a' 2. We then get a' 2 = aU 2 + b' with Ha" 2"A2
~
1/4, K(t,b;A 1,A 4 )
~
e
C· (tl2) • And so on, for ever. This
e
shows that K(t,a;A 1,A 4 ) ~ C· t lIallA2 and A2 c (A 1,A4 ) • The desired conclusion, viz A2 = fA l ,A4 )e (= 1/2 o?~olff's theorem) follows now by invoking the rei teration theo~m (§ 8).# Wolff's theorem has many applications. Example. SUP!\OSEZo we have in one way or other estabY.shed that, on the one hand, (H ,L)e = LP and, on the other hand, (L ,BMO) = LP (cf. the discussion in 6, example 1). Then re can, usin~ WoHr, fuge these two results and conclude that indeed (H ,BMO)e = L , all this in the natural hypothesis on the parameters, viz. lip = rE e.
f
Mh. Now we have summarized some of the most pertinent facts about K (and J) spaces (the equivalence theorem, the duality theorem and the 'rei teration theorem, § § 5, 7, 8) and we have also seen some 11 ve specimens of such fellows (§ 6). But we have not yet accounted for ,their interpolation property, which is really the heart of the matter. It Is high time to remedy this! Let us begin with the very simplest result. Let A = (AO,A ) and B = (BO'B l ) be any two compatible pairs of Banach spaces. If lj. is a linear map from AO + A1 into BO + Bl such that the restriction T I Ai of T to the space Ai defines a cont!.nuo~s linear map from Ai into B., this for i = 0,1, we agree to write T : A + B (or T : AO + ~O' Al + Bl )·l Similarly,if A and B are intermediate spaces, for A and B respectively, and the restriction T IA of T to A maps ~ continuously into B, we write T: A + B. If always (for any T) T : A + B implies T : A + B one says that ~ and!
s.
e
< 1, 1 S q S .. are relative interpolation
174
J. PEETRE
INTERPOLATION THEOREM. T: 1, 1 ::s q ::s CIII.
A+
-
-
B impUes T : Aeq + Beq' 0 < e <
Indeed, this an almost immediate consequence of the inequality
where MO and M.1 are the norms of the restrictions T : AO + BO and T : Al + Bl respectIvely. For (6) gives
where at the last step one makes a change of variable in the integral involved. So if Aeq is equipped with the norm laUA eq = +eq(K(t,a;A»
-
-
and Be with the norm UbHB eq = +eq(K(t,b;B» we arrive at the "convexCfty " inequality (7)
1'1 < 1'I 1 -
0
el'le
l'
M being the norm of T : Aa + Be ' which is even more than required for the theorem. In such a si~uationqone says that one has (eaxct) interpolation of power e. The proof of (6) likewise follows from first principles: If a = a O + a 1 with a. E A. (i = 0,1) then Ta = TaO + Tal with now Ta. E B. (i = 0,1). Therefbre wL conclude that 1 1
=
Taking the infimum over all possible such decompositions of a, the desired formula (6) results.# Example. Parseval implies Paley. On any (commutative) group G we have
f: rf: L2 where
L1 + LCIII trivial £orm o£ Riemann-Lebesgue
+
r!
Parseval (or Plancherel? or Pythagoras??)
stands for the Fourier transform. So by interpolation ~
V-:
1
2
CIII
2
(L , L ) eq + (L , L ) eq·
Now by our previous information (§ 6) (Ll,L2)
q
= LPq, (Loo,Ll)e
q
=
PARACOMMUTATORS AND MINIMAL SPACES
175
LP',q if lip = 1 - a, lip' = a, i.e. IIp + IIp' = 1. Tberefore we bave Paley's tbeorem: LP,q + LP',q if IIp + IIp'
= 1,
Since LPq is monotone in tbe argument q, Hausdorff-Ti tcbmarsh-Riesz):
f:
LP + LP '
if IIp + IIp'
= 1, P
P
e
(1 ,2).
tbis impUes (Young-
e
(1,2).
!.L.. Next we state an important complement to the Interpolation theorem, concerning interpolation of bilinear (and, more generally, multilinear) operators. Let th!re thus be gi v!n three compatible pairs of Banach spaces A = (AO'A l ), B =(B O,B l ) and C =(CO'C l ). If T is a bilinear ~ap ~efi~ed in (AO + At) X (B O + Bt ) witb values in Co + C l we write T : A B + C to mean that the restrictlon T 1Ai Bi of
,
T to AL Bi defines a continuous linear map from Ai X ~ into C i , tbis for i = 1),1. In a similar way we u!e !,he_ notation T : A X B" + C where A, B, C are interpolation spaces for A, B, C respectively. BILINEAR INTERPOLATION THEOREM. T : A B + C implies T: Aa Bar + caw where 0 < a < 1, 1 < q, r, w < e» and lIw = l/q + llr - 1, q We shall not give the proof. Let us only pOint out tbat tbe crucial 'relation llw = lIq + llr - 1 (which, incidentally, puts some restrictions 'on the applicability of the theorem) comes from Young's inequality for 'convolutions. Note in particlliar th.!lt we can take q = 1, w = r so that in particular (r = e») T : Aal X Bae» + Cae» is a valid conclusion. Example. Let B be any Banach space continously imbedded as a dense Subspace in a Hilbert space. Then we are allowed to form the "self-dual" compati~e pair (B,B*) (both Band H are imbedded in the cont,ining space B ). At least if B is reflexive one can show that (B.B )112,2 =H, a most useful result for the applications. (Notice also 1 e» 1 ,.the essentially "self-dual" nature of (L ,L ). But, unfortunately, L is "hot a reflexive space.)
Remark. Both the bilinear interpolation ,theorem and the * above example extend to the complex method. Thus 10 particular [B,B ] 112 = H. Also we dispense with all this fuzz with the parameters. ~ Returning to the interpolation of linear operators we notice that f.he convexit¥ inequality (7) may be interpreted as saying that the Space L(A ,B ) is "half" of class a with respect to the operator pair . aq aq ,l 'CL(AO,BO)L(A l ,B l ». (Here L(X, Y) denotes the spaces of all continuous
J. PEETRE
176
linear operators from a Banach space X into another Y.) In particular, we may draw the conclusion that
for a E (0,1), 1 ~ q following useful result.
~
co. Now we supplement this formula by the
PROPOSITION. (L(A O,BO),L(A 1,B 1»aco
c::
-
-
L(A a1 ,LB aco )'
The proof goes via bilinear interpolation. Consider the map (a, T) ~ Ta (where T is a generic operator from A into 8, not the T in the theorem). Then we have
whence by a special case of the bilinear interpolation theorem (see the comments after that theorem)
which is what we set out to prove.# Remark. Stein's classical interpolation theorem is essentially a result of the same nature (with complex interpolation). The usefulness of this result depends on the circumstance that one often
a
* f.
Then the conclusion is that T: L P ~
Lq for l/q = lip - a/n (p > n/a). Naively, one would like to approach this via Young's inequality for convolutions (f * gELq if f E LP , g E Lr, l/q = lip + llr - 1>. In o~r case lIr = 1 - a/n but the function ka narrowly escapes to be of class Lr. Instead it lies in the "Marcinkiewicz space" Lrco • So it is in the interpolation space (L r O,L r 1>BI and we have a corresponding standard decomposition k = 1: k . (see th': a
aJ
definition of J spac!f' § 4); actually, one can take k Ii = restriction of k to the set {RJ < I x I < RJ} for an approprrate choice of R. ~rrespondingly we have a decomposition of the o~erator T, T = 1: Tj , and we may infer that T E (L(LPO,L~),L(LPl,L U.s." with suitabl'e exponents P9' qo' Pl' qt' So our proposition is applical)le.
PARACOMMUTATORS AND MINIMAL SPACES
177
We invite the reader to work out the details of the above sketch. (Else consult myoid paper [P21 which contains many more illustrations of this simple but surprisingly powerful approach.) One can also prove a sharpening of Young's inequality itself. Let us briefly indicate this, phrasing the details of the proof in a slightly different way. ExamDle. O'Neil's inequality: f E LPe , g E Lroo => f * gELqoo provided IIp = lIq + lIr - 1, no special restriction on e. For the proof let us write Young's inequality twice:
and interpolate: p
(L 0, L
p
1)
91
v
A
r r (L O,L 1)900
~
q q (L O,L 1)900.
If we work this out we obtain
Now we interpolate once more (keeping r fixed!) and the desired result follows. #
178
J. PEETRE
Notes for Lecture 1. There are now several excellent books (monographs) available which are in whole or in part devoted to the theory of interpolation spaces. Besides [Bll let us mention [Bl0], [ K3] , [ T2], [B91. There exists also a rather complete bibliography of interpolation spaces (until the year 1980> [C3]. Among more recent works of survey character let us mention [Oll, [03], [P5] (see also [P6]). The theory of interpolation spaces took its birth in the years around 1960 and among its fathers we mention Aronszajn, Lions, Gagliardo, Calderon, Krern. It has its origin in the classical interpolation theorems of Riesz-Thorin <1926, 1939) and Marcinkiewicz (1939), For a full account of the latter we refer to chap. 12 of Zygmund's book [Z]. There also Stein's theorem is discussed. It is interesting to note that part of the motivation for the study of "abstract" interpolation came from problems about Sobolev spaces (~ainly the scale of spaces HS(Q», that is, Sobolev spaces based on the L metric) then current in elliptic POE. See the treatise by Lions and Magenes [L5] where one major tool is precisely the interpolation property of the scale HS(m. Later on interpolation spaces have found applications in a manifold of branches of Analysis: harmonic analysis, approximation theory, geometry of Banach spaces. Back to POE! In my opinion, in later years the theory of interpolation spaces has developed in a far too abstract direction, getting more and more divorced from the rest of Analysis. This is no good sign. Perhaps it is still time to stem the tide. In particular, it should be of major interest to see what the new developmemts associated with the "new" generation: Ov~innikov, Janson, Brudnyl' and Kruglyak, and their followers, really means for the down to the earth analysis problems from which interpolation once arose. It is not my intention to write a history of interpolation here, but let me trace at least some of the most important developments. The complex method was simul taneousl y designed by several mathematicians but is usually associated with the name of Calderon and this with considerable right. Indeed, Calderon at once gave a fullfledged theory and his paper [Cl] is still the paper par excellence on the topic of complex interpolation. Very little genuinely new has evolved here later, if we exclude the more recent extension to infinitely many spaces (see e.g. [C7] and the references given there; for an extension in a different direction cf. also [R5]). The complex theory is to some extent parallel to the real theory, as outlined here. The real method too has its origin in the work of several mathematicians. Let me especially stress Gagliardo's contributions (see e.g. [Gll). Thus the K functional is at least implicit there. It was later independentl y rediscovered by Oklander (see [01]) and others. The main body of theorems about K and J spaces as presented here (equivalence theorem, duality theorem, reiteration theorem) is due to Lions and can in principle be found in [L6]. These ideas were then taken up by the Butzer school and applied in the context of approximati;j.. th
PARACOMMUTATORS AND MINIMAL SPACES
179
groundbreaking work of BrudnyT and Krugl yak and their associates (Asekritova, BereznoYetc.), gravitating around the strong form of the fundamental lemma, first proved by these authors [ B8]; the problem seems to have been open for many years. Other proofs of the fundamental lemma can be found in [03], [C9] and elsewhere. The formal notion of scale was conceived by Krein around 1960 (see [K3], chap. 3). It has received a renewed interest recently through Wolff's theorem [W] (§ 8); see the discussion in [J4]. In 1965 Aronszajn and Gagliardo published in important theorem [A6] where the abstract foundations of interpolation are laid. In particular, they proved the existence of minimal and maximal interpolation spaces Un a technical sense). At that epoque the paper arose no great interest. Its importance was regognized only some 10 years later, towards the end of the 70's, the credit going to the Vorone~ (Krefn) school (see the survey article [0]); especially, I would like to emphasize the name Ov~innikov. But the real breakthrough came only with Janson's paper [J31. It has been followed up by others. Let me mention here the thesis of Nilsson [ N2] , [ N3] (summarized in [P5]). Thanks to all these authors, we now know that (almost) all interesting interpolation methods (functors) arise as minimal or maximal spaces ("orbits" and "corbits") in the sense of Aronszajn ,nd Gagliardo [ A6]. Thus the J spaces come as orbi ts oJ weighted L pairs and, dually, the K spaces as orbits of weighted L pairs. Likewise the complex spaces admit a (co)orbit description. Besides the "classical" interpolation methods (real and complex, as we met them here) there is a third type of interpolation functors. One gets them .by taking coorbits of weighted L pairs and orbits of weighted L pairs - Hand G in the notation of [J3]. For instance, the G method is essentiall y the ± method originating from Gagliardo [G3] and then used in interpolation of Orlicz and other r.i. spaces in [02], [ G5]. The idea is roughl y to use unconditional, rather than absolute summation in the definition, whence ±. A problem, mostly of technical nature, which has attracted much attention is the problem of describing all compatible pairs of Banach spaces A =(AQ'A ) such that all interpolation spaces for A are K spaces (with a general lunction parameter .). This problem has its origin in the work of Calder6n [C2] and Mityagin [M3], who gave the example A (A ,A 1). Accordingly one speaks of Calder6n(-Mityagin) pairs. Despi~e quite a many papers devoted to this subject (see in particular the discussion in [03], [P5]; as an example of a recent contribution we list [Cl0]) it appears that we are quite far from its solution.
=
180
J. PEETRE
LECTURE 2. BESOV SPACES. 1. In this lecture we shall be concerned with a special scale of
sp~es, namely Besov space Bsq •
Historicall y speaking, BeJ6v spaces arose out of Sobolev spaces (thus again in the context of ellipti~ POE). As is well-known, Sobolev back in the 30's defined the spaces W (0) for 1 S P S 00, k an integer> 0 p
rtt
an open set in Rn (definition: f E Wk(O)~v a E (multi-index) Oaf E LP(O) (deriv~rt'tves in distributPon sense). We was eS!kecially interested in the restriction (the "trace") of functions in W (0) to manifolds of lower dimension, say, a smooth piece r of the b8undary aQ. It turns out that full answer is in terms of the Besov spaces HSq(r>. More precisely, a function g on r is the trace of a function f in W lQ) P lIp k
2. The guiding idea is to break up a given function (or distribution) f many pieces such that the Fourier transform of the \I th piece f is, roughly speaking, concentrated near the sphere {~ E Rn : I~ I = 2"} (\I E Z). This is known as the (a) dyadic decomposition
in
i!: into countabl y
~f.
.,....
N.B. - Strictly speaking, we have to deal with two spaces': in duality. We will not distinguish them in notation but we usually reserve Greek letters for the dual (phase) variables (~,1'), ••• ) using Latin ones (x,y, ••• ) for elements in the "direct" (configuration) space. The duality will be denoted by <x,~> = x1~1 + ••• + xn~n. The Fourier transform is (formally) given by A
£(~)
=
J e i<x
l<> £(x)
,~
dx.
....,
Rn
We also write Iff =
i.
In praxis, such a decomposition of f can be gotten by first picking a sequence of test functions {cp } ~ such taht each cp is in f (the v:v n:v Schwartz class of rapidly decreasin unctions in.a.) witH
PARACOMMUTATORS AND MINIMAL SPACES (~
-
181
n E R •
01
n EN). .".. 1\
We then put f = cp * f (* = convolution). Clearly supp f e: U too for each v E Z ~nd, l'f we also requir" that (*) 1: cp Yx) =v6 (delta function) or, equivalently, (**) 1:VE~ cpv(~) = 1 we 'h«iie aVtrue partition: f = 1: vEZ fv. However, it is sometimes not suitable to have such a strong conortion as (*) or (**) and then we have instead to impose a Tauberian condition m5rmander-Shapiro):
3. We can now define the Besov space Bsq = p real, 0 < p, q ~ 00:
rfpq (r:l1), ~
s arbitrary
Here {fv> vEZ is any dyadic partition associated with f and fO = 1:~Of. va isVeasy to see that Bsq is in general a quasi-Banach space (for the natural choice of the Pquasi -norm) and we have the continuous imbeddings :Ie: Bsq e:J' (tempered distributions). It is a Banach space if p ~ 1, q ~ 1 ada for simplicity we will in what follows usually make this assumption. (Sometimes it is wise to exclude the extremal cases p or q = 1 too.) Otherwise, the case 0 < p < 1 is quite subtle since then the space LP cannot be realized as a space of distributions (theorem of Day). 4. For technical reasons it is often convenient (and sometimes necessary) to use a related scale of spaces, namely the so-called q homogeneous Besov spaces (by contrast to the previous non-homogeneous ones). The idecfis to let the dyadic partition run all the way down to -00:
as
£ E
bsq ~){~ p
~Z
(2 vS R£V UL P )q}l/ q < 00.
w-J
The price we must pay in order to get (quasi-)Banach spaces here is that we have to reckon modulo polynomials. More precisely, let N be an int~er > s - nip (~ s - nip if we impose the restriction q = 1>. Then Ffq can be realized as a Banach space modulo pol ynomials of degree < N¥ (If N = 0 this means no extra qualification.) We say "homogeneous" because the spaces rfq turn out to be dilation p ...sq def ...sq invariant: If f E ts p then f). (x) = f()'x) E tsp and moreover
J. PEETRE
182
By contrast, the spaces Bsq do not possess this nice property. (Needless to point out, Ffq ~d all other spaces we deal with here are translation invariant too.~ Example. 6 (delta function) E B-n/p,oo for all 1 ~ p ~ p
00.
Indeed, if f =
6 so that "f(~) :: 1 we have a dyadic partition {f } EZ of f such that IIfvllLP l¥ 2n p. By duality (see § 5) it follows that '{.,evliWVe the imbedding Bn/p,l l¥ C (continuous functions). Indeed, if 1 ~ p ~ 2 one hase even a Iltronger result in the form of the factorization:
This is essentially the Bernstein-Szasz theorem about absolutely convergent Fourier transforms. There is something special wi th the value s =nip! 5. As a first result on Besov spaces we mention the DUALITY THEOREM. (Bsq )* l¥ b-s:-q ' where' stands for the conjugate exponent; here we assumJ> p < 00 aRd q < 00. This is proven by some juggling with Hahn-Banach (alternatively one could have used the duality theorem for abstract interpolation; see § 8 where the link with interpolation is established). The (anti-)dua~ty e~ployed is of course the one coming from the inner product in L = L (Rn ): .",
(f,g) ~
J f(x)g(x)
--
dx.
Rn
Remark. One can also include the limiting case p = 00 or q = 00 if one replaces q by the closure of in the corresponding metric. (If p < 00 and q < 00 p is automatically dense in Ffq, as is readily seen.) Notice that the space Ffoo for s non-integer >PO is the well-known Lipschitz space (defined usT'ng first order differences), whereas for s integer it is the Zygmund ~ace (defined using second order differences). One often writes A = rf (even for s ~ 0>, and" for the corresponding closure s 00 s * -sl -sl* of j. Thus we have in particular "s l¥ Bl ' , (B l ' ) .. As so As is the second dual of" (a fact ignored by many). s
es
PARACOMMUTATORS AND MINIMAL SPACES
183
6. As we have already mentioned, a major problem in the theory of Sobolev and Besov spaces, which has attracted the attention of many workers (especially in the Soviet UnionD, is the characterization of the trace on smooth submanifolds. If we restrict attention to the limiting case when the dimension is precisely n, we have the Sobolev imbedding problem (that is, the study of relations between spaces with different parameters in any of our scales). Here is the result for (homogeneous) Besov spaces. IMBEDDING THEOREM. We have
provided s ~ sl' P S P1' q S q1 and (that is the important thing!> s nip = sl - n/p1· Let us indicate the proof in the special case q =q =00. Consider any d~adic decomposition {f } E~ a given function (distribution) f in Eroo , say, f = (f) * f whef1e )' r.} EZ is a family of p v v vv_
0\
test functions of the type considered in § 2. Let
{(f) v}
-
-
vEZ be another A
such sequence, "slightly larger" than (ell v } vEZ in the sense that eII(Z;) A
1 in the support of I'
v
(f)
v
•
=
Then we have the reproducing property
d~I ell v * I
=
(f)
v
*
(f)
v
*
I = ell v
*
Iv.
From Young's inequality for convolution (see Lecture 1) we have
Now it is easy to see that 1Ie11 IIL P = O(2vn<1-1I~) and moreover, using finally that f E Ffoo, we have Irf IILP = O(2vs ). From this again we infer that p v
As {f' } E is an equally good dyadic decomposition for f, Concl ucfe ~ii-
we
J. PEETRE
184
That we get such a transparent proof is of course a sign of that we have picked the "right" definition of our spaces.! 7. Besides Besov spaces let us also mention "potential" spaces. To be on the safe side we assume presently that 1 < P < GO. One possible definition is
where {f } E again is a suitable dyadic partiti0ll, of f. EquAvalently, in terms ~f ~~Ipotential" operator IS (formally: (If) (~) = I ~ I f(~»:
It follows (apely the Mihlin multiplier theorem!) that, provided s is an integer ~ 0, P'" = 9f ("homogeneous" Sobolev space: fEW s # rf f E
P
P
P
LP for all a E Nn with I a I = s). Limiting case (s = 0>: pO = LP • It is -
p
likewise elementary (Parseval> to see that ~ = 9~2. This is typical for the case p = 2. In the general case (p" 2) one can only prove inclusions:
nPsp
c:
ps
P
c:
nPs2
(1 < P < 2) ,
ns2 p
c:
p P c: P
nPsp
(2 < P <
GO).
This is quite deep and essentially the only proof goes via PaleyLittlewood theory. Indeed, by the latter
from which the above imbeddings readily follow. Remarks. 1. This suggests the introduction of the even more general spaces known as Triebel-Lizorkin spaces:
They have the virtue of connecting in a better way with potential (and Sobolev) spaces than Besov spaces but have not (yet?) found that widespread use in the applications. We refer to the works of Triebel (e.g. his most recent book [T3]). 2. Notice that all the spaces rJ& for p fixed (as well as rfq for p, q fixed) are canonically isomorplRc. the operator IS proJiding this canonical isomorphism. If 1 < p, q < GO they are as Banach spaces isomorphic to LP (0.1) (or IPU q ». regardless of the dimension n.
185
PARACOMMUTATORS AND MINIMAL SPACES
8. Let us now look at our spaces from the interpolation point of view. The simplest result in this direction is (1)
*
for s = <1 - 9)s + 9s , 0 < 9 < 1, So sl (no relation between the q'sD From the imbecl>ding theorems (§ 7), in conjunction with the reiteration theorem, it follows now that likewise
(1 ' ) for s = <1 - 9)sO + 9s1' 0 < 9 < 1, So * sl. In principle, results like (1) or <1') are comprised in Lions's theory of interpolation of semi group spaces (mentioned in Lecture 1, § 6), modulo the finite difference description (see § 9). However, it is easy to give a direct proof based on the dyadic decomposition. Let us indicate the argument. Let f E esq be given. If {f } EZ is the dyadic decomposition p
v
v """
associated with f, let us write f = f' + fll with f' = 1: f fll - 1: v~N v' - v
= C.2 N(SO-S)·U£U Bsq p
and analogously (if sl > s)
n£"u Bs l l S C·2 N(s -S)·n£UBsq. p
p
Thus we may concl ude that
if s =
(1 -
9)sO +
Bsq P
c
aS l
(taking t
a
l1li
2
N(s -s 1 0). This proves
(BsO l BS l 1 ) • p' P 9-
Similarly, using the J functional instead of the K functional. we may establish that
J. PEETRE
186
( hBO W hB 9
Dp
IDp1
hBq.
)91 c: Dp
The desired conclusion, viz. identity (1), now follows from the imbeddings just established by application of the reiteration theorem (§ 8 of Lecture 1). One can also prove that
*
for s = (1 - 9)sO + 9s , s sl' l/q = (1 - 9)/qO + 9/ql' 0 < 9 < 1. Notice that here it is essen\:iaP to fiave the right connection between the q's. By contrast, the scale of potential spaces is stable for complex interpolation:
*
if s = (1 - 9)sO + 9s , So sl' 0 < 9 < 1. I! we also let lihe metric LP, vary the situation becomes more involved. Thus the three parameter family sq is not any longer stable for real interpolation and the resulting irRerpolation spaces have a very complicated description (see [P91). Essentially only in one
e
*
if s = (1 - 9)sO + 9s 1, So sl' lip = (1 - 9)/PO + 9/Pl' 0 < 9 < 1. Thus the parameters must be adjus1.ed to each other.
Remark. These and other basic results on interpolation of Sobolev and Besov spaces have their origin in the thesis of Grisvard [G4 1, whom, as we have already remarked, one should never forget to quote. As a Besov s~ace Efq in a natural way may be realized as a "retract" of a space IS (LP ) (~eighted lq with values in LP) - this goes via the dyadic decomposition, of course - they have a natural interpretation in terms of the interpolation theory of vector valued LP (see Lecture 1, § 6). 9. Besides the road followed here (the method of dyadic partitions) there are many approaches to Besov space theory. For instance, the results of § 8 lend themselves to a interpolation definition of Efq (that is, we take formula (1) as definition). A related definition, prefJi.red by (S.M.) Nikol'skir and followers (see e.g. [B3], [Nll), goes via approximation theory, more precisely, using the best approximation by entire functions of exponential type (analogues of the theorems of Jackson and Bernstein). One can also employ spline and other similar type of functions (especially in local problems).
PARACOMMUTATORS AND MINIMAL SPACES
187
However, there are two classical procedures, which both eminate from Hardy and Littlewood (compare the Notes for this Lecture). One is via finite differences, the other goes via harmonic continuation (Poisson transform). Indeed, they may both be treated on equal footing within a more general framework, which incidentally contains the previous dyadic formalism too. Let us brien y indicate what this is about! (Not too many people seem to be aware of these connections!) For greater clarity we prefer now to use a "continuous" variable t (t > Q) in place of the previous "discrete" one \I (\I E l:) hitherto favorized (connection: t l1li 2\1). Let there be given a "kernel" k = k(x) inlr-n. For t > 0 set k = k t (x) = t -nk(x/t). The idea is roughly speaking (still using "dyadic" ~anguage) that one in the definition of the dyadic partition (see § 2) relaxes the condition of the support of the Fourier transform of the test functions cp to be confined to the seXs U but keeps the Tauberian condi tion. T~at is, now reverting to k, k(~) "must not vanish in U but must satisfy some growth conditions at 0 and at 00. Then one has ~e following "metatheorem":
We shall not enter into the details, which may be found in [P2], chap. 8. Instead we sh~ll mention some illustrative examples. Let us " however point out that, very roughly, one needs that k(~) is O( I~ I s ), s or perhaps better o( I ~ I ), in some sense. For simplicity, let us presently confine attention to the case n = 1. "
.~
Examples. 1. Finite differences. We take k(~) = (e 1 -, integer> o. Then one is lead to expect a formula of the type
N
U , N an
(2)
N
N-1
Here tat = tatta and tatf = f(x + t) - f(x). But in which range of s can we hope to be valid conclusion? By our previous "criterion", we must have
(ei~ - UN = O( I~ IN). This gives 0 ~ s ~ N. Indeed, a closer examination reveals that (2) is true without any further qualifications if only 0 < s < N. (To remove any doubts on this point, I suggest you verify (2) directly in the special case p = q = 2, by a brute (but instructive) calculation, just use Plancherel.) One can also prove such a result: Write s > 0 in the form s = k + 01, k integer, 0 < a ~ 1. Then
188
If P
J. PEETRE
=q =
00
we recognize here the definition of the Zygmund class A • a
" 2. Harmonic continuation. k(x) = lin· 11<1 + x 2), k(~) = exp(- I~ 1>. Now u = u t = k t * f(x), viewed as a function of x and t, is harmonic . 2 n+1 . (6 lU = Q) 10 the upper halfplane R (upper halfspace l.. 1f n > 1); thn 1S why we say harmonic continu~ion. In this case we expect that
The requirement e - I ~ I = O( I ~ IS) gives s S 0 and, indeed, (3) turns out to be rigorously true if s < o. To do better we have to take derivatives. For instance, for s < 1, one can prove that
And so on for higher deri vati ves. Remark. Instead of the Poisson transform one ~n ~s well take ~he Gauss-Weierstrass transform (k(x) = 1I2,/n· exp(-x ), k(~) = exp(-~ "». Therefore our spaces also admit a description i~ ter~ of "temperatures" (solutions of the heat equation aulat = a u/ax ), rather than harmonic functions. 10*• The general question treated in § 9 can be given yet another twist. Let a "convolution kernel" k be given and set as before kt(x) =
= t -nk(x/t)
0h
(t > so that " k t (~) " = k(t~). If f is a given function (distribution) in.a,., set also u = k t * f; u will be though~+ff as a function of two variables. thus living In the upper halfspace.R:. n If X is any (quasi-)normed ~ace of functions or distribUtIons in l.and ~ a positive measure on we are interested in the inequality
t:
More generally, if X is only semi-(quasi-)normed but the semi(quasi-)norm induces a (quasi-)norm modulo polynomials of degree < a fixed integer N, it is natural to take instead the inequality
PARACOMMUT ATORS AND MINIMAL SPACES
(X,p,n)
{~
J IOau(x,t) IP
189
d~}l/P
la'r=N Rn+1
S cUfU X for f E X.
~
Let I be a cu~e in ~+1 of side length 1 = 1(1), with one of its faces contained in $-. We say that ~ is d-dimensioanl Un the sense of Carleson) if for all I holds the inequality (C)
~
(I)
S Cl
d
•
Example. Carleson measures in the ordinary sense have dimension n. The following two lemmata are left as an excercise for the reader. LEMMA 1. An equivalent condition, for d > n, is (C*)
~(I*) S C*ld
where 1* is the "upper" half of 1.# ooLEMMcf 2. Let f be a po:itive f~ctio~ of one variable such that dtlt < 00. Then J 0 f('; I x I + t'" d~ < 00 for d-dimensional ,measures. #
J0 f(Ut
One can now prove the following THEOREM. The following conditions, where we wri te X = Ssp and, similarly, use XL when s.p are replaced by sl' Pl' are equivalent? (i) ~ is (X,p)-Carleson for !!2m!!. reasonable kernel k. Ui) Same thing for all ditto. (Ui) ~ is (X1 ,Pl)-Carleson for all reasonable kernels k and any PI with sl = s + n(I/ P l - lip). Uv) ~ is d-dimensional where d = n - pes - N). What "reasonable" is will be disclosed in the proof (growth condition + mean value property). Proof. (after [HS1) For simplicity let N = o. (1) ~ (U>. Logic. (11) ~ U11>. Just apply the imbedding X c: Xl. (Ui) ~
(~(I»l/P S ClfllX
S Cl n/P l- s 1lfl 1
Xl
= Cl n/p - s = Cl d/P •
190
J. PEETRE
(iv) =>*(1). p > 1. At this stage at last "resonable" is needed. Let 1** denote I blown up in some fixed scale, say 11: 10. We require now the following MEAN VALUE PROPERTY: lu(x,t) I
~ C{t-nJ lul P dXdt/t}l/ P for (x,t)
E
I*.
I** Take this to power p and integrate: lu(x,t) I P ~ Cpt-nJ
lul P dxdt/t for (x,t) E
I**
I** J
I*
lul P dl-l ~ Cptd-nJ
lul P dxdt/t ~
I**
cP J
(t - 9 lu I)P dxdt/t,
I**
where we also used inequality (C *). With now loss of generality we may assume that I-l has compact support contained in a preascribed cube I. Now we make the same decomposition of I as in the proof of Lemma 1. (Surely you have done your home work!> We get the same inequality
. sets Iv* and Iv ** wlth bounded overlap of tne I
(.r* v
= 1,2, ••• ). Finally sum over v. Because of 's we get
But if we impose proper GROWTH CONDITIONS on our kernel
(see the
*
previous §) nf n d~f (J (t -s I k t f I P dxdtlt) lip is an equivalent norm in X. p ~ 1. This is easy. Under suitable GROWTH CONDITIONS on k we get p
lip
from Lemma 2: (J I k I dl-l) is as a p-normed a space is ("minimality"):
d-n Ca for a > O. On the other handd K generated by the functions k a -n a
~
any f E X has the representation f = 1: X k a n - d where 1: I X IP < 00.# v v av v
asP
is the largest space among all spaces COROLLARY. If d > n then X which have the d-dimeRsional measures as (X,N,p)-Carleson measures.# Remark. An even simpler way of doing it: Postulate (as a substitute for the mean value property) on the one hand that
for the functions u = k ... * f, on the other hand that
191
PARACOMMUTATORS AND MINIMAL SPACES
f
A(
X
+
t
S "C -, t) dl·d X,
t)
~
Ct
d
for d-dimensional measures ~. (Here A is a suitable positive kernel.) Then there is NOTHING to prove; in particular, the "geometry" gets eliminated. 11*. We discuss briefly the mean value property encountered in § 10. Gauss observed that if u is any harmonic function (A u = 0) then its value at a point P can be recaptured from its average over either a sphere S or a ball B about P: u(P) = 111 S I • f s....U = 11 I B I • f BU. This can be generalized to more general (hypo-)ellipticr>DE"2 To be specific, consider biharmonic equatio~ A u = O. Let G be the corresponding Green's function with pole P (A G = cS p ) and Dirichlet's boundary condition (G IS = BGIBn IS = 0; n is the exterior normal). Using twice Green's formula (1 B (ltv - uAv) 3v = 1S (BulBn e v -
uBv/Bn ) dS) I get u(P) = f (BAGIBn u - AGBulBn ) dV. Next apply this for4fuula to a family of c8ncentric b\lls B with rldii r between two r fixed numbers and form the average. Then I get a formula of the type u(P) = 1A au + 1A bBulBr where a and bare smooth functions supported by an "annulus" ~ about P, BIBr denoting radial derivation. By partial integration I get rid of the derivative BulBr and there results a formula of the same type with b = 0.# This procedure clearly is quite general. In partic1:f.\Sr, it applies to dilation invariant hypoelliptic equations in ~ of the form P(t,BIBt,BIBx)u = O. There results a mean value theorem of the desired type: lu(x,t) I ~ Ct-nf
lu(Z;,"C) I dZ;d"CI"C for (x,t) E 1*.
1** with I * and I ** as before (§ 10). Examples. 1) B2uIBt2 + Au = 0 (Laplace's equation). 2) 112t· BulBt - Au = 0 (heat equation in slight disguise). 3) B4 ulBt4 + 2B 2AulBt2 + A2u = 0 (biharmonic equation; the case we reall y have discussed here).
2 2 4) B ulBt + (2q + O/t· BulBt + Au = 0 Uhe singular GASPT equation of Alexander Weinstein; see e.g. Lions's (1 ) book [L31, chap. XII). In all these examples we can apply the result to the solution u of an appropriate boundary problem with boundary datum f given on~. This solution apparently is of the type u =,.kt * f. Below I list the Fourier transform k of k in each of these cases.
,.
1)
k(Z;) = exp(- I Z; I) (corresponding to the Poisson kernel).
192
J. PEETRE
2) k(~) kernel).
=
exp(-I~ 1~ (corresponding to the Gauss-Weierstrass
A
3)
k(~)
=
exp(-I~
1)(1 -
I~ I).
k(~) = const I (I ~ 1)/1 ~ 1q
the modified Bessel function; if q = q Apparently, the argument just given is confined to hypo-elliptic equations. 4)
-1/2 one gets back Sase 1».
2 2 2 2 Counterexample. The wave equation B u/Bx - B u/By = 0, which has unbounded integrable solutions. Ouestion. Is the mean value property true also for other kernels k, not necessarily connected with a (hypo)elliptic POE? 12. Most of what we have done in the foregoing belongs to "soft" Analysis. But one virtue of this is that it easily generalizes to more general situations. For instance, in the definition of Besov spaces (§ 3,4) one can replace the underlying space LP = LP(~) by practically any translation invariant space X of functions or (tempered) distributions in Rn. One then gets the spaces BsqX, esqx, in the notation of [P2]. For instance, an important case is when X is a multiplier space, say, X = (Fourier) mul tipliers on LP itself; see again [P2]. One can also replace the parameters s,q by a general "function parameter" (cf. Lecture 1, § 5, remark). In this context, let us quote Janson's paper [J2]. Another possibility, popular in the Soviet Union. especially the Moscow area, is to allow different homogeneity in different directions; one is then lead to anisotropic Besov (and Sobolev) spaces (see [B4], [Nll).
There is also a pure I y abstract direction. Let X be any Banach space and assume that the (additive) group Rn acts on X by sufficiently n smooth operators. Let the group operator correspoll~ing to x E,&: be denoted U (so that U = U U , Uo = id, U = (U) ). If cp is any x x+y x y -x x sufficiently well behaved function living on Rn we define the "convolution" cp )(. f with elements f of X: cp X f
=
f
cp(x)U x dx.
Rn This allows to define "abstract;' Besov space: We say that
for some (any) dyadic partition (cp ) vEZ (cf. § 2). Much of the previous developments can now be ca~ri8dtibver to this general case; we refer to [P2l, chap. 10.
PARACOMMUTATORS AND MINIMAL SPACES
193
13. * Examples. 1. If X is a translation invariant function (or distribution) space over Rn and U is taken to be the usual action of Rn on X
(the "Rudin ball") in ~ (or rather its boundary the sphere S = S2n-l). Indeed. the complex geometry of S provides us with a natural action of the one dimensional group R. Let z = (zl ••••• z ) be a point of S. Then in the (real) tangent space T S at that poin¥ we distinguish several interesting directions. On thZe one hand. we have the complex tangent vectors making up the complex tangent space TCS = T SniT S at z. On the other hand. we have the complex normal iz~ defin~ng a ~ector field on S. This vector field provides us with the desired grou~ jlctio~. Its effect on a function f on Sis: f(z) gets replaced by
f
Ji, and
Spaces of Besov type for the complex unit ball and even more general domains (of the Siegel type) have been considered by Hahn and Mi tchell (see [ Hll) and others. Probably one can also construct a decent theory for general strictly pseudo-convex domains. Let us mention the work of Stein. Krantz etc. (see e.g. [Kll. [S5]). 4. (after an idea of Per Nilsson's) Lastly. one can define in a natural way Besov spaces of operators in a Hilbert space H. say. If U is a unitary group of operators in H we have severa~ interesting acti~ns on operators. for instance A ~ U AU and A ~ U AU • Assume. taking n = x x x x 1. that there is an o.n. basis e k such that Ux: e k
~ eikXek (k = 1.2•••• ).
Then the "convolution" corresPWlds for an operator ~ with matrix (a ) to Schur multiplication by
194
J. PEETRE
Notes for Lecture 2. There are many books on Sobolev and Besov spaces. Besides [P2] let us also mention [All, [56], [Nll, [B4], [T21. In the case of one variable and the circle T there is already a theory of Besov spaces in slight disguise in the fundamental work of Hardy, Littlewood and Paley around 1930. Most of the relevant papers can be found in part III of Hardy's Collected Papers [H21. In particular, the dyadic partition is there. As for myself I learnt it about from H6rmander who at an early stage employed it systematically at least in the special case p = q = 2, of paramount importance in (linear) POE; see [H3], p. 50 Somewhat later H.S. Shapiro comes into the picture approaching the whole subject from the point of view of approximation theory; see e.g. [53]. Also NikolskiY's outlook was basically approximation theoretic; see the books [ Nll, [B4] already quoted. Oleg Besov got his name attached to these spaces, because he wrote a great memoir [B2] about them in 1960. Somewhat later another equally important paper was written in the U.S., the thesis of Taibleson [Til. It is also fair, in this connection, to record Gagliardo's beautiful paper [Gll from 1957, because it has been so influential, providing the key to the trace problem. Nowadays there exists also a theory of Sobolev and Besov spaces for general subsets of ~ (not just smooth submanifolds) developed by Wallin and his followers in UmeA (see [J8]); this may loosely be thought of as an LP version of the classical Whitney extension theorem (see e.g. [56]). Parallel work has been done by Brudnyi and others (see e.g. [B7]). In this case the present Fourier approach of course fails and has to be replaced by the "oldfashioned" finite difference techniques. For a recent "atomic" d.escription of Besov spaces see [F5]. As for the extension to the case 0 < p < 1 we emphasize in particular the pioneering work of Flett (see e.g. [F3]). Among later contributors we mention here the name Triebel (see the treatise [T2] already quoted and the most recent [T3]). Triebel was in particular interested in proving a priori estimates for elliptic P.D.E. (of the Agmon-DouglisNirenberg type) for the whole range p E (O,GO] and recently this goal has been achieved in all generality by a brilliant student of his by name of Franke. As for Triebel-Lizorkin spaces a comprehensivce treatment can be found in [T31. Here, as in the 0 < P < 1 theory of Besov spaces, the g~eat Fefferman-Stein paper [F2] on "real" HP spaces (and the dual of H , not to forget) turned out to be seminal.
PARACOMMUTATORS AND MINIMAL SPACES
195
LECTURE 3. MOBIUS INVARIANT FUNCTION SPACES.
h We have now developed lots of machinery (Lectures 1 and 2). Of course, only a tiny part of it will be used effectively in the sequel but I think it is good to have this general background in mind. In the present Lecture we shall turn to a theme which I was lead to through contacts with Jonathan Arazy and Stephen Fisher. (Some of what I am going to say is though joint work with Svante Janson; see especially § 4 and 7.) This, quite generally, is the study of Banach spaces of functions which are invariant under a given group 0 acting in a prescribed manner on the functions (maybe, in part, via its action on the underlying space). Remark. As usual, quasi-Banach spaces are equally good in most cases. As an example we may quote the rearrangement invariant (r.i.) spaces on a given measure space (0 = measure preserving transformations ("change of variable") + change of sign of the function). Here we shall however be concerned wi th cases when 0 is a semi -simple Lie group with non-trivial center and the simplest example is then 0 = the M5bius group of (conformal> automorphisms of either the unit disk U in the complex plane s;.,or, which, by conformal equivalence, essentially amounts to the same, the upper (Poincar") halfplane P; we are in the first place interested in the natural action (via "change of variable") of this group on holomorphic functions. Whereas all r.t. spaces in a way can be reconstructed, from two GO extremal ones, Land L (this is essentially the Calderon-Mityagin theorem; see Notes for Lecture 1>, the situation for M5bius invariant Banach spaces of holomorphic functions is much more complex. There is surely an abundance of such spaces, some of them quite important in Analysis, but, at the present stage, it is far from clear what the "general" Mobius invariant space might be like. For previous treatments of the theory we refer to [A4] and [ P4] , where other material complementary to what we say here may be found too, and more references. 2. Let us first make precise the terminology. U is the (open) unit disk { I z I < I} in the complex plane S' .I. = a U is its boundary, the unit circumference { I z I = 1}. o ("M5bius group") is the group of all conformal selfmaps (IIM5bius transformations") of U or P; it can be realized a~ the group of real unimodular 2 X 2 matrices SL(2,R) modulo ±id; if (arf) E SL(2,R) (where -by c.lZ = thus ad - bc = 1> it induces 'IN'!' the M5bius map c.l c 0 given (az + b)/(cz + d) (z E P); in the U realization we have instead the subgroup SUU,1> of SL(2,,,. o acts on (holomorphic) functions defined on U via f(z) -+ f(c.lz) (c.l E 0); sometimes it is fonvenient to put C f = f 0 c.l. Clear: C 0 C = C"c.l' C id = id, C~ = Cc.l -1. So, stric'tly speaking, C is c.lan aGtirepresenf"ation, not a representation.
J. PEETRE
196
Remark. Most of what we do here extends formall y to the more general case of the "multiplier" (anti-)representation
~ is given, usually an integer (to maintain the one valuedness). This is connected with what is known as the holomorphic discrete series; for details see [P3].
If X is a (quasi-)Banach space of holomorphic functions in U we say that X (isometrically) M5bius invariant if f E X ~ IC fa E X «(,) E G) and in addition (this is essential!> IC fn = nfn (f E X, Ii E G). Thus G acls on X via isometries. (,) There is also the more general case when this latter assumption is dropped so that G acts on X only via isomorphisms (bounded operators). Then we may say that X is isomorphically M5bius invariant but this case will not be treated here. We will likewise consider the case when our space is a Banch space only modulo constants. So there are really two theories, one dealing with true invariant norms, the other with invariant semi-norms. If X is a space "of type I" then, forgetting the constants, one gets from it the "type II" space X/C, but the process cannot be reversed in general (see [A5]). With a little extra care no confusion will ever arise.
Remark. In the case of multiplier representations with ~ integer> 1 (see previous remark) one similarly has to reckon modulo polynomials of degree < ~; ~ = 1 is the primi ti ve case we started out wi tho Let also, once for all. point out that when dealing with spaces of holomorphic functions over U we al ways identify an element of our space with it distributional boundary values (provided the latter exist, which they do indeed in the cases of interest to us). Finall y, in the one dimensional case (by contrast to the higher dimensional case, see § 7-8) the use of holomorphic functions is not that crucial. Indeed, most of what we do extends mutatis muntandis to harmonic functions. 3. Examples. 1. The Dirichlet space«). f EO.>~J J U I f'(z) 12 do < • (dO =It.H.~ormalized area measure). It can be realized as a Besov space: D = B2 ' (T) if, adopting the convention just made (§ 2), we identify holomorphic functions with their boundary values. Clearly ff) is M5bius invariant: An easy computation reveals that Bfa
=
{f f
If'
(z)
12 do }1/2
U
is an invariant semi-norm (modulo constants). The simplest proof of the identification with a Besov space depends on Parseval's formula:
PARACOMMUTATORS AND MINIMAL SPACES
197
where "f(n) are the Fourier (Taylor) coefficients of f. Another uselful expression (out of may equivalent ones) for the norm: nfll =
[f I Ifn:'~l)
TXT .,,;.,.,.
-
f(~2)
12
Id~ll Id~2
II
1~1
-
~212
]1/2,
up to normalization. The important thing is that(1) is a Hilbert space and that f comes from an inner product. FACT 1 ([A4]). (Jis the only MObius invariant Hilbert space of holomorphic functions over U. Indeed, the inner product int1)may be used as an invariant duality for general MObius invariant spaces (Banach, not Hilbert). Thus by polarization we obtain from any of our expressions for the invariant norm an invariant (anti-)duality:
II
f' (z)g' (z) da,
U
~O
IJ(f(~1)-f(~2»(g(~1)-g(~2» Id~11Id~21/1~1- ~212. l~
Another useful expression (perhaps the most natural one) is:
Thus the entire theory is essentially self-dual. 2. As a generalization of the Dirichlet space alone has the spaces~, 1 < P S 00:
~II If(~l) - f(~2) IPld~ll Id~21/1~1 - ~212 <
00
lXI,
Obviously these expres'}y,n~ correspond to invariant (semi-)norms. It is readily seen that~ = Bp , (ih so we have again a Besov space. 1 1 Likewise, the space Bl (.1' i. MO~J.lIIIinvariant. We therefore
.At PUtDl
def =
J. PEETRE
198
B~,I. It is also natural to set fA d~f B lIp,P for 0 < p < 1; this is a quasi -Banach space, not a Bana~bPspace. I!'rom the theory of Besov spaces (Lecture 2, § 2) we see thatlj* _Il " 1 ~ P < duality just defined. p ""l>
10,
in the invariant
FACT 2 ([ A4])."1 is the minimal MObius invariant Banach space; for every other such space X one has the continuous imbedding~ c X. It follows thataJt is the II_hull of G: every f ~ 031 can be written as a sum f = 1: A.w. witll w. E G, 1: 1Ai 1 < 10. (For p-normed spaces (with p fixed, 0 < pI 1) one has a similar result, with the role of the minimal space enjoyed by the space 8; see [CG].) Similarly:
<-
p
FACT 3 ([R4]). The space (Banach) space: X c 11) for any X.
43
is the maximal MObius invariant
10
10
This is rather simple so we may as well indicate the proof. In the above definition in the limiting case p = 10 only the first definition is any good and we interpret it then as: £ E
63 ~ sup 10
(1
-
1z 12) 1£' (z) 1 <
10.
zEU
Thus/B is nothing but the famous Bloch space, from Bloch's theorem (see e.g. [A3]; see also the survey [C4]). It is clear from this that the space surely is MObius invariant, for the condition may be written as sup E 1(f 0 w)'(O) 1 < 10. But for any X one has the inequality 1f'(O) 1 ~ C 11111, E X. (This is one of our taci t assumptions about our spaces of holomorphic functions; something has to be continuous!> Now apply this to the functions f 0 w , w E G, and the desired imbedding follows.# In the previous discussion we worked "modulo constants". If we require genuine norms, the supply of ~aces gets more restricted. In particular, the maximal space is then H • The minimal space is still~, for~ can be equipped with an invariant norm; take, for instance, II fll = II fll 10 + II fU' where UfU' is any of the invariant semi-norms. The two 10 spaces are of course not any longer each others duals. But H possesses an interesting MObius invariant pre-dual, which can be described as follows:
q
4. Another interesting Mobius invariant space is BMO. Here is a definition (due to Garcia) which is automatically MObius invariant: £ E BMO~)sup zEU
I
.... T
I£(~) - £(z) mz(~)
Id~1 <
10,
PARACOMMUTATORS AND MINIMAL SPACES
199
where n (l;) is the Poisson ke~nel. In other ~ords, SMO is the M6bius "coorbiti! of the Hardy space H •
where e sup goes) over all finite families of disjoint arcs (ak,bk ) of Not much ems t e known about these spaces.
J.
4*. How to find ones way in this wilderness? In this Section we briefly indicate one possibility to organize the previous rather incoherent material, while a different approach will be suggested in the following one. Our idea is to associate various families of M6bius invariant spaces wi th the different orbits under the action of a on the product set clos U clos U. It turns out that most known spaces fall under this heading. In some cases one likewise gets alternative descriptions of them. Let thus the M6bius group a act on clos U X clos U via (zl,z2) .. (wz 1,wz2 ). Let fl denote the diagonal, fl = {(zl,z2) e clos U X clos U : zl = z2}; It may be identified with clos U itself: fi. - clos U. On U U we consider the invariant distance d(zl,z2) = IZ 1 - Z21/Jl'l1l'l2. (Quite
.
2
generally, we write now l'l = 1 - Iz I for z e U.) Here is a complete list of the possible types of orbits under this action: 1)
u> nfl- U.
2) (TXI) n fl- T. 3) Any subset E;, say, of \fl defined by the condition d(zl,z2) = ", " a constant o. (E" may be identified with the set of hyperl50hc segements of length ".) 4)
(1)( I) \
fl.
5) Any of the two subsets U XI and I)( U. Each of these spaces can be implemented as a homogeneous space, that is, in the form a/H where H is a suitable subgroup of a. Further, if H is unimodular a/H admits an invariant measure, because a itself is unimodular. This occurs in the cases 1) and 3-5) but not in crse 2). 1) al {rotations}. The invariant (Poincart§) measure is dOl = do/l'l. 2) al {dialations + translations}. (To see this, pass "["0 the upper halfplane realization, P instead of U. The transformations which then fix the point at infinity GO are of the form z .. az + b (a, b real, b .. 0>.) Of course, there exists no dilation + translation invariant measure. 3) a itself. (A geometric argument shows that a acts simply transitively on each set E".) An invariant measure is gotten by transplanting the Haar measure on a to E". 4) a/{dilations}. (The maps which fix GO and one more point, say, the origin 0 are of t~e form z .. az (a real .. 0).) The invariant measure is an old friend: do i = I dl;l I I dl;2 1/1l;1 - l;2 I •
200
J. PEETRE
5) Again G itself. (The only transformation which fixes one point at the boundary, the pOint -1, say, and ~e an interior point 0, say, is the identity.) The invariant measure dO i is again gotten from the Haar measure. We describe now a general principle for obtaining invariant function spaces. Let 0 be any measure space on which an action of G is defined (usually via measure preserving transformations). Let further T be a map which to holomorphic functions on U associates measurable functions on 0 and in addition "intertwi!\fs" the two group actions (the one on 0 and the one on U>. Denote by T the formal adjoint of T (with respect to the natural G invariant pairings on U and 0 respectiv~ly, induced by the inner products in the Dirichlet space 6l and in L (0) respecti~ely). In the cases of interest to us T will be an isofetry from into L (0). Therefore we will havt, formally at least, ToT = id. This means
Here z and z2 are any two filled pOints in U with d(z ,z2) = " and we integra,"e witli respect to the Haar measure d6) on G. tase 4) is quite similar to the cases 1) and 3). For example, we get now back, from a more general point of view, one of the characterizations of (f:J in § 3. Finally, Garcia's definition of BMO (Ukewise § 3) obviously is connected with case 5).
201
PARACOMMUTATORS AND MINIMAL SPACES
5*. In this Section we discuss the questionrP.iven an invariant space X what can be said about the sequence z I X} >0 of norms of the "basis" {zn} n>O. n
{n
Example. If X =~ then IIznll SIll n and if X then IIznll the general case we expect zn to be somewhere in't,etween.
=13
SIll
1, so in
We introduce the quantity x(m,k)
dEf sup
I
wEG
m k z ,z >1 w
(=
X (k,
m) )
where we thus sup over all "matrix elements" of the representation C, <.,. > being the M~bius invariant inner product induced by the norm of the Dirichlet space (§ 3). Remark. Indeed, the matrix elements can be expressed in terms of hypergeometric functions (alternatively Jacobi polynomials). They also satisfy a 2nd order linear DE. (lowe these two pieces of information to T. Koornwinder and U. Haagerup respectively.) One can do the same game with the more general actions C Il (see the corresponding remark in § 2). THEOREM. 1) Let X be a G invariant Banch space of holomorphic functions on U. Then
2) Conversely, if H(k) is a positive function on the positive integers (a "sequence") such that (**)
H(m)x(m,k)
~
mH(k),
then there exist a G invariant space X as above with Proof. This (t)
m liz II
(where II· II
i~-view
* \=
of the following general formula
m m/llz II
* ist the norm dual to
m k I
w
II z k II = H(k).
~
m liz II
II •
II).
* liz k II
Therefore k m = mllz II/liz II,
whence the desired relati~ (*) by the definition of X. 2) Define X to be the 1 -hull of the elements C zm /H(m). It is clear that IIzmll ~ H(m). For the converse we use aga~n formula (t). Quite generally holds
202
J. PEETRE
I I
I<£,e z m>I/H(m), w
sup
m,w so by
(**)
I j
we obtain
m
k
supl
m
II
supx(m,k) k
~
m/H(In)j
1 = H(m).#
=q then (*) gives (as now Ilzmll ,. m) x(m,k) ~ cm or, in view of
the symmetry, x(m,k) ~ c min(m,k). Thus (within equivalence) if (**)'
~
H(m) min(l,k/m)
~
(**)
is in particular fulfilled
cH(k).
Therefore we get COROLLARY. If H(m) is equivalent to a concave function then there exists a G-invariant space X such that IIzmll ,. H(m).# This again leads to the question whether there is an inequality in the opposite direction: x(m,k) ~ c'min(m,k) (c' > 0>. However, a somewhat heuristic calculation of Haagerup's based on the WKB (or GreenLiouville) approximation points to that this is not so. So maybe there are "exotic" M6bius invariant spaces which cannot be obtained by "interpolation" even in a very weak sense. 6. As an application of the previous theory, more precisel y, the idea of the minimal space (see § 3), let us know sketch a simple proof of the trace ideal criterion for Hankel operators, usually associated with the name of Peller (see [P6]), in the case 1 ~ P < 00. (The case 0 < P < 1 is much harder; see [P7], [S11.) It is expedient to consider Hankel forms, rather than operators. By a Hankel form with symbol
r
(£, g) =
f
~£g Idz 1/2n
2 (£,gEH).
;t..
The name of Hankel is motivated by the fact that on the "canonical" basis (zn) we have
=
A
cp(n + m),
so r is associated with a Hankel matrix. The question is now when is of'litrace (Schatten-von Neumann) class SP, 1 < P < 00.
r
1
203
PARACOMMUTATORS AND MINIMAL SPACES
Remark. Generally speaking, a bilinear form r over the product H}C K of two Hilbert sp'aces Hand K is of class sP if some associated operator T is of class Sp. (We say that T : H .. K is assocaited with r if r<x,y) =
Ur..
implies that the norm + .. IIS 1 is G-invariant. The analogue for C- 1 of the Arazy-Fisher theorem ("A51 (see § 3) then gives
or, using the above relation between the two "symbols" cp and +,
This is one endpoint for the interpolation. The other endpoint is gotten from Nehari's theorem
Consequently, the interpolation gives half of Peller's theorem; the converse follows by a duality argument (see [P61, [P3]).' Remark. The prime virtue of the above argument is that it can be adapted to a variety of similar situations (typical abstract nonsense). For instance, it extends to Hankel forms in several variables (cf. next §) and to "paracommutators" in Rn (cf. Lecture 4). Notice that in the latter case the group under consideration is not any longer semisimple. Indeed, even in the present case (unit disk) invariance under the full Mabius group is not that important; what is essential is invariance just under its "parabolic" subgroup (dilations + translations).
1...:.. * Most of what we have been doing until now extends mutatis mutandis to the case of holomorphic functions over the unit ball B = B in n, But there is one initial difficulty. Naively, one would 2think o"l the following analogue of the Dirichlet integral: ~ II df(z).. do. (z), where II. stands for the invariant norm . duce~ _ by the lBerdman metric on tile cotangent bundle of B {explicitly: df Ui 222 .. (1 - II z. )( adf a - I Rf I ) where a· II is the corresponding Euclidean norm, Rf =1: z.8f/8z i being the radial derivative of f) and do. is the 1 1
r.
a·
=.
J. PEETRE
204
2 n+1
corresponding volume element (explicitely: da i = datU - UzU) , da being the Euclidean ditto). The drawback is that for f holomorphic the integral is divergent (unless f == const). So one has to proceed differently. One possibility is first to consider the "norm"
J 1f ( z) 12 (1
-
II z H2 ) a
da ( z ) ,
B
a a complex parameter, which is invariant if we let f transform according to a suitable multiplier representation, and subsequently to use analytic continuation; this is the technique of M. Riesz. N. B. - "Invariant" in the above discussion refers, of course, to the group of holomorphic selfmaps of B, which is isomorphic to the projective group PU
IB
(0)
~(K)
a
71)'
S Ca,
where K is the (geodesic) ball of radius a about (Bergma~ metric
Zo in the invariant
Consider the map Tf(z) = (f(z) -. f(zo»/d(z,zO)' where d(z,zO) is the separation between z and Zo in that metric. If f EIB (Bloch space) then 00
00
It follows that ( 1)
PARACOMMUTATORS AND MINIMAL SPACES
205
On the other handS1 c: HOD. Thus if f E ~1 If' (z)
C211 f' 1118 •
f' (zO) I ~
-
1
Consider the set
E~
where I Tf(z) I !!:
Then d(z,zO) ~ C211 fll CC211 f IIISt /~ and (2 )
1/~
fed
=
0
so
~,
that is, I fez) - f(zo) I
E~ c:
~
d(z,zO).
Ko. Thus (0) gives
IJ(E~)
~
T:!II.... L 1 k ( dlJ ) • -t wea
Interpolation between
(1)
and
(2)
gives
We have proved: THEOREM. If
(0)
is fulfilled then
Remark. Actually, the proof gives more: the same conclusion with replaced by (H OD )1/2.2. Unfortunately, nobody knows what the latter space is (cf. the ":'elected problems" in [P3]).
.8..
The question is now if there are any interesting measures satisfying condition (0). One can always take d to be arc length along a geodesic issuing from zOo But consider V = B n]in (the "real" unit ball). Then V is a maximal totally geodesic submanifold of B. It seems therefore natural to take (now Zo = 0, supp IJ c: V) dlJ(x)
=
(d(x,O»
1-n
d"t'i -
Ixl
1-n
(1 -
2 -1
Ixl)
d-r
Where d't" and d't". are the corresponding "real" volume elements, the Euclidean and thd invariant one. There results the following FejilrRiesz type inequality:
Where we make the normalization f(O) := O. This is the result in [S3] if n • 1.
J. PEETRE
206
Remark. The general problem is about measures on B
JJ (I£(z)
B such that
- £(w) I/d(z, w»p dv(z, w)
BXB Call such a measure a generalized ~ ,p)-Carleson measure. The result just proven has thus the general patt~n: generalized weak(~,l)-Carleson ~ generalized ctt,p)-Carleson (p > 1>. At!, ,p)-Carleson measure (cf. Lecture 2, § 10) corresponds then to the li~illng case when all mass sits on the diagonal.
L* The ultimate setting for the theory is perhaps something like a Siegel domain. Here 2 we content ourselves with a few words ab~ut the case of the bidisk U = UX U = «z,w) : Iz 1 < 1, Iw 1 < 1} in £.. (The formally more general case of the polydisk is of course quite analogous.) The group 02' say, of all automorphlms of U2 is generated by transformations of the following two types: 10 the maps
where each of the "partial" maps z -+ (a l z + b 1)/(c l z + d ) and w -+ (a~~ + b 2)/(c2w + d 2 ) belongs to the previous "l-dlmensional J group 0. ""2 the map (z,w) -+ (w,z) ("symmetry"). We inquire2 whether there exist Hilbert spaces of holomorphic functions on U which are invariant under 02. Here is the result:
'8
THEOREM. the only 02-invariant Hilbert spaces of holomorphic functions on U are the following ones: 10 41= (1) 20 1tconsists of all holomorphic functions f of the type f = fl (z) + f 2 (w), with
"',s,:
U£U
2
=
~ L..
A 2 ~ A 2 n 1£1 (n) 1 + m 1£2(m) 1 < n=1 m=1
L..
00;
to get a genuine Hilbert space we have to work modulo constants, as usual. 3°1(consists of all holomorphic functions f wi th
If 12
A = ~ nm I£(n, m) 12 < n= ,m=1
or, equivalently,
00
207
PARACOMMUTATORS AND MINIMAL SPACES
bow we have to reckon modulo functions of the type f~z) + f 2(w).# The interesting case, apparently, is only the case"""3°. In "this case we can likewise define a maximal space and a minimal one. In particular, the maximal space ("Bloch space") is defined by the condition sup (1 z,v
Iz I
2
) (1 -
Iv I
2
2
) I a f (z, v) I azav I <
01.
J. PEETRE
208
LECTURE 4. P ARACOMMUT ATOR5.
L. In Lecture 3 we encountered the trace ideal criterion for Hankel operators and we saw that invariance considerations, in particular the idea of the minimal space, were most useful in that context. Now it is well-known (see e.g. [R21) that, in one variable, the study of Hankel operators to some extent is equivalent to the study of commutators of / Calderon-Zygmund operators. (By commutator we mean in this connection always the commutator with a multiplication operator.) Namely, on the real line 1. (as we have asserted, whether we consider! or 1. is not terribly important in the present situation, mainly because of conformal invariance) there is essenly only one Calderbn-Zygmund o~erator, viz. the Hilbert transform H (Hf = lin· p. v. llx * f). We have H = id
n
,
2. We put ourselves in Ji . By a Calderon-Zygmund operator we mean an operator of the form Tf = p.v. k * f where k = k(x) is a kernel homogeneous of degree -n with 0 spherical means: I 5"k(x)" d5 = 0 for any sphere about the origin; the Fourier transform k = k(~) is then homogeneous of degree O.
" Example. k(~) = any of i~./I ~ I (j = 1, ... ,n). The corresponding operators T are the famous Rie~ transforms and are usually denoted by Ri (j = 1, ••• ,n). If n = 1 there is only one Riesz transform and this is the HUbert transform H (§ 1). The following result is classical (see e.g. [541). THEOREM (Calder6n-Zygmund if n > 1; M. Riesz if n = 1). If 1 < P < then T extends to a continuous map from LP into LP •
GO
But we are more interested in the commutators [T,bl where as before (§ 1) we identify the function b = b(x) in £n with the corresponding multiplication operator; explici tel y: [T,b If = k * bf b(k * f). The first results on such commutators were obtained by Coifman, Rochberg and Weiss in their celebrated paper [C61, essentially the special case when T is a Riesz transform, and their work was then completed by Uchi yama [U] and by Janson [Jl].
PARACOMMUTATORS AND MINIMAL SPACES
THEOREM. [T,b] is bounded in LP , 1 < P <
209 00,
iff b e BMO.
Remark. To avoid tri viali ties we assume throughout that k ;: O. Uchiyama [U] further settles the question when [T,bl is a compact operator: the condition b e BMO has then to be replaced by b e CMO (continuous, not bounded mean oscillation). The next step was taken by Janson and Wolff [J7 1, who considered the question when a commutator is of trace class. Of course, we ar~ now concerned with the action ofd~rr s$Paces on the Hilbert space L only. In what follows we write B = B P, s = nIp (cf. Lecture 2 for the definition of the "homogeneous" IPesov s8aces). THEOREM. Let n > 1, 0 < P < 00. Then, if p > n, [T,b] is in sP iff b e ~. If P S n, then [T,b] cannot be in sP unless b ;: const. This last part of the theorem indicates a quite new phenomenon not present in the one dimensional case. ~ Let us indicate the main features of the proof of the Janson-Wolff theorem (§ 2). We begin by the observation that [T,b] is an integral operator with kernel k(x - yHb(x) - b(y». Now, if f ell it is easy to see - this is, for a change, an instance we we have to ~esort to the finite difference charayterization of Besov space that (b(x) - b(y»/1 x - yin e LP(LP 00). On the other hand, one has the following result, a "real" version of Russo's "complex" theorem [R31.
LEMMA ([ J7]). Let V be any integrrl operator with kernel j(x, y). If j(x,y':>, as well as its "transpose" t1 (x,y) = j(y,x) both belong to LP(L P 00), where p > 2, then V is in sP (Lorentz trace class). If we combine this lemma with the above estimate we see that [T,b] e Spoo, provided b e ' , p > n. Another interpolation sharpens this to [T,b] esP, so we haviproved the "easy" half of the theorem. In [J7] the converse ([T,b] e sP ~ b eB, , p> n) is much harder and resides on a description of Besov spaces uiing mean oscillation over a mesh of cubes. Luckily there is a much simpler approach and this depends on a broadening of the whole setup. Besides the previous operators [T,bl we consider also higher order commutators [T1 [ T 2 ••• [T ,b] .•• ] 1, where the T i again are Calder6n-Zygmund Operators, wrtth kernels k i • Then we are able to bypass the unpleasant "barrier" p = N and are free to use the "minimal" spaceB1! Before making this extension let us say a few words about the case p ~ N in the original Janson-Wolff theorem. The simplest way to see why such an obstacle towards the commutator being of trace class comes up Is to pass to the outlook of 900 (pseudo differential operators).
J. PEETRE
210
~ ak(~)/a~ .• ab(x)/aX .• ~1
J
J
By sui table cutoffs we mad' further reduce oneself to a compact situation, say, the torus T • Quite generally, a YDO A, say, with - T n is in S p we must have m > nIp, unless symbol a(x,~) of order -m on a(x,~) :: O. This can be seen by "freezing the coefficients"
L* Now I briefly sketch work done in last summer ('83) while Svanle Janson visited Lund for a couple of days. For details see [J5]. We have thus N Calder6n-Zygmund operators T 1, ••• ,T N in ~ with kernel k 1, ••• ,kN and we form the N fold commutator with multiplication with the functlon b = b(x):
The first issue on our agenda is the question of boundedness. Of cours~ it is clear that if b E BMO the new commutator r is still bounded in L (and in LP , 1 < P < 00, too, for that matter). The nice thing that also the converse is true. THEOREM 1. r
THnOREM 2. r
t&o.
e~1
Next we turn to our m~in question, that of trace class commutators. We consider r acting on L •
6.
2 0 If P THEOREM 3. 10 If P > min(n/N,1) then r 1 then rN.
~
N.B. - The above formulation is somewhat inprecise, because we require also certain non-degeneracy conditions on the set of kernels k 1, ... ,k N, to be explained below. It is clear that this result contains the Janson-Wolff theorem (the case N = 1), see § 2-3. Surprisingly, the proof becomes simpler and in a way much more natural, rendering the difficult converse (§ 3) almost trivial. Let us see whv this is so.
PARACOMMUTATORS AND MINIMAL SPACES
211
We begin with the direct part. p > 2. This portion of the proof is patterned on the corresponding part of the proof in the Janson-Wolff theorem (see § 3). We begin with the observation that r(b) is an integral operator whose kernel is given by the following BEAUTIFUL formula:
I(
-1 ) N
J... J
k 1 (z 1 ) ... kN ( zN ) t. z 1· .. t.ZN b do
I
where t. b(x) d~f b(x) - b(x - z) and the "(N - I)-fold" ~ntegral is taken over thl affine manifold zl + ••• + z = x - y in (R n ) • So everything blows down to proving the !Dclusion n
ap b E Bp=>I x -
YI
n(1-1/p)-aJ
• ••
J
IZ 1 I
-n
• •• I zN I
-n
•
the rest is exactly as in the case N = 1. Again, to prove this inclusion it suffices to do it in the endpoint cases p = 1 and p = 00 and then to interpolate. The case p = 00 is rather straightforward, while for p = 1 it suffices to check the thing for one particulf\r "test function" b in~ This because of the minimality of the space e~ for the fOlloW}n~ action of .the group of dilations on functions b in Rn : b(x) ~ 6 n b(6x) (cf. ·!Lecture 2, § 3). (This relat~ to the con:rderations of Lecture 3 as follows: The minimality of B1 (I) among all Mobius invariant Banach :, spaces of holomorphic functions in the unit disk U does, on a closer examination, not utilize the full Mobius group but only its "parabolic" subgroup, the group of translations + dilations, if we use the halfplane language.) The proof of the la~ter fact is based on simple "Tauberian" considerations: The j\space e~ is "generated" by functions b the :Fouurier transforms b having their supports in the "dyadic" seY.s U V v ( see Lecture 2, § 2). 1 ~ It is here that the full power of the minimal property of e~ is used in a more essential way (the special case a = 1). Namely, in order to prove the boundedness of ~(b) for b E and N > nit suffices to do ,this for one single b. This gives the desirea result in the endpoint case ., = 1. To get it for intermediate values of p one uses complex interpolation, to be exact interpolation of operators a la Stein. We notice baht the "Fourier kernel" corresponding to the previous kernel has the equally BEAUTIFUL expression:
4\
By polarization we may take kl = ••• = kN = k, in which case the formula ,-ecluces to
212
J. PEETRE
The complex operator family used in [34] now corresponds to the famil y of Fourier kernels:
Unfortunately, what is not quite that simple there is the verification of the boundedness of r(b) for a test function bi there is no good reason for why this should be so. We come now to the converse. We shall be quite brief on this since it is based on the same type of duality argument as in Peller's case [P6] (cf. Lecture 3, § 6). If we thus have a result of the type r : ~ ~ sP we p
0
get a result of the type r*: sP ~ for the adjoint r* of r. Here in forming adjoints we use the ordinar}p (anti-) duality (b,c) ~ J b(\)c(X) dx for functions b, c in Rn and t~ iuality (T,S) ~ trace TS for operators T, S in the Hilbert siace L (R ). It is now essential to know that the composite map ToT, which'""'is a convolution operator, is sufficiently close to the identity. This requires a non-degeneracy condition on k 1, ••. ,kN and in [J4] the following one is used:
Until now we have only discussed the "favorable" case p > n/N, i.e. the case when there is a large supply of commutators in Sp. If on the other hand p :s: N, we require a stronger condition, viz.
If thus this condition is fulfilled, the conclusion of part 20 of theorem 3
is valid. From the !PDQ point of view, hRwe~er, thAt most natural condition is in terms of the Poisson brackets (k 1, (k2, ••• (kN,b} ••• } }. Explicitely:
~
b is a polynomial of degree < N.
Thus we have in toto three conditions. It is easy to see that (***) ~ ~ (*). (For instance, (*~ follows from (***) if we in the latter condition plug in b(x) = <e,x> (a homogeneous polynomial of degree!).) Somewhat surprisingly, however, these conditions are not equivalent, as simple counterexamples reveal. (**)
PARACOMMUTATORS AND MINIMAL SPACES
213
* Lastly, we come to the topic which has motivated the title of the ~ present Lecture. The methods and ideas used in the theory we have been discussing in previous § were of quite general nature but they were used in an after all rather special si tuation, viz. the "higher" commutators. But already in the proof we had on at least one occasion to leave this realm, namely while considering the complex family of operators with Fourier kernels cp (~,'1). So it is natural to ask if there is not a larger domain to whicli the same type of analysis applies. Indeed, this is the case and we shall now report very briefl y on this new extension of Hankel (and Toeplitz) theory. Again it question of (unpublished) joint work with Svante Janson; see also [T4]. We have baptized these new awful beasts "paracommutators", because of some formal resemblance with the paramultiplication of Bony - a highly popular theme in certain circles right now (see e.g. [B5]; see also the even more "popular" account by Strichartz [S7]). It is convenient to consider them as bilinear forms, rather than operators. To be exact, by a paracommutator we shall mean a bilinear form of the type
If
(f, g) = (2n)-n
b(~ +
A
A
'1)A(~,'1)f(~)g('1)
d~d'1
_ n Rl)cR
IW
where b = b(x) is a given function in Rn, the "symbol" of r, while A(~,'1) is a (Fourier) kernel, usually fixed throughout the discussion. To indicate the b dependence one writes r = reb). If A(~,'1) == 1 then r reduces to the form J b(-x)f(x)g(x) dx. Thus, in the general case, r may be conceived as a modification of this "multiplication" form via the Schur multiplier A(~,'1). In this way the theory of Hankel and Toeplitz operators gets connected with the vast domain of Schur multiplication. It is pertinent to mention in this context the work of Birman and Solomyak (see e.g. [B4]). Another advantage of the "bilinear" outlook is that one is immediately lead to multilinear generalizations, that is, analogues of r with more than two function arguments. (More on multilinear forms in the following Lecture!) Examples. 1. Let n = 1 and take A(~,'1) = 1 if ~'1 > 0, 0 else. Then we have essentially a Toeplitz form (operator). 2. Similarly, if still n = 1 but A(~,'1) = 1 if ~'1 < 0, 0 else we are in the Hankel si tuation. " - k('1), " k" the Fourier transform of a 3. For general n, if A(~,'1) = k(~) Calder6n-Zygmund kernel k =k(x), we are back in the situation of § 2. 4. More generall y, taking several factors, we get the "higher" commutators discussed in § 4. We now report on progress made on extending the previous theory to the case of paracommutators, i.e., the general body of results dealing with the interplay between a symbol b and the form r = r b. To formulate our main result (analogue of Jh, trace"ideal c~iterion) we require the "Varopoulos algebra" W(X)C Y) ! LGO(X) eL GO(y), where X and Yare arbitrary (measurable) subsetsGOof Rn. (A E W(XGO X Y)~A(~,'1) = I: "'jbj(~)cl'1) where I: I"'j I < GO, IbjilL (X) :!; 1, IICjllL (Y) :!; 1. Norm:
214
J. PEETRE
IIAU = inf I: I"jl.) THEOREM. Assume: i) A E W(R'kR n ), ii> II A(~, T) II W(K X K) ~ C - K of radius r about the point Ir, I) n for some N > 0 and every- ball (r,,-r,), r, E!,n arbitrary. Then if p > nlN we have b E~ ~ r
(rl
For 1 ~ P < 00 it also possible to prove a converse result (by the same duality argument as before). It is conceivable that the same conclusion remains in force also for 0 < p < 1 and that the method in [ P7] (or the one in [S1]) will do it. Finally, we have also the obvious analogue of theorem 1 (boundedness) and theorem 2 (Bloch symbols) in § 4.
PARACOMMUTATORS AND MINIMAL SPACES
215
LECTURE 5. MULTILINEAR FORMS.
L. In the previous discussion (see Lectures 3 and 4) we have seen that in dealing with Hankel and similar operators it is expedient also to put into play the corresponding bilinear form. One virture of this bilinear philosophy is that one is naturally lead to ask to what extent the whole theory can be generalized to multilinear forms as well.
A(x, y, z •••• ) =
.
~
~1
"i <x, ei>
1
Where {e , ••• ,e }. {fl ••••• f } •••• are orthonormal systems of vectors in en and th~" i cSmp.i.ex numlYers.
J. PEETRE
216
If r = 2 then all (bilinear) forms are S-forms. If r > 2 this is not the case.
Indeed, a count on ones fingers quickly reveals that the space of all S-forms of given rank topologically has dimension rn(n - 1) + 2r. On the r other hand, the vector space of all rank rr tensors is 2n •
2n r > rn(n - 1) + 2r if r > 2,
n > 1.
Therefore if r > 2 not all forms are S-forms. Thus Schmidt forms are no good if we wish to get a "normal" form for general r-forms (r > 2). Remark. The name of Schmidt is, historically speaking, perhaps not entirely justified, because in the finite dimensioanl case such representations of bilinear forms (operators) were considered earlier by Beltrami and Jordan. (Again I learnt this from Ake BjOrk.) 3*. We suggest now a possible way of getting a different normal form for r-tensors, r > 2. Imitating the usual procedure in the case r = 2 we associate with a given form A an extremal problem. For simplicity we take r = 3. Let thus A = A(x,y,z) be given. To maximize IA(x,y,z) I under the side conditions UxH = Uyll = Hzll = 1. Let x = e 1, y = f 1, z = gl be a stationary (critical> point for this extremal problem. A vector x close to e 1 can be written as x = e 1 + ex' + ••• with x, .1. e 1• This gives (derfVation with respect to e!f A(x',f 1,gl) = O. In the same way we obtain A(e 1,y',gl) = 0 if y' .1. f A(e 1,ft!z'T = 0 if Z' .1. gl' If we now for general x, y, z write x = ae 1 x', y =of 1 + y', Z = cg 1 + z, (with x, .1. e 1 etc.) we get
t
+ bA ( x' , f l' z')
+ cA ( x' , y' , 9 1 ) + A ( x ' , y , , z' )
(with 5 terms instead of expected 8D. A is thus determined by the following data: 1) a complex number A(e 1,f 1,gl) * 0 (unless A == 0>, which can be normalized to be > O. 2) three bilinear forms A(e 1,·,·), A(·,f 1,·), A(·,·,gl) in n-l variables. 3) a trilinear form a' = A(: , • , .) in n - 1 variables. 4) three unit vectors e 1, f 1, gl' (A r9ugh check shows :that toe n~mber of parameters is the correct one: 2n = 1 + 3' 2' (n - 1) + 2(n - 1) + 3(2n - 1) - 2; the last term (with the minus sign) accounts for the fact that et! f 1, gl are not uniquely determined; it is permitted to multiply them oy unimodular numbers 01, ~, Y with OI~y= 1.> Now repeat the process with A replaced by A'. By induction we then get three canonical orthonormal bases {ei } , {fi } , {gi } associated with A.
PARACOMMUTATORS AND MINIMAL SPACES
217
For general mul tUinear forms A one simUarl y has the following "normal" form (multi = r): A
1
A (x, y) =
~
B , X ,
1=1
1
1
Y,
(Bi
1
~
B
J, if i
< j).
Remark. Here is a problem which I have not been able to solve, We have only dealt with (more or less unique?) maximum of the modulus IA(x,y,z) I, but what about all the other stationary points? In particular, what about their number?
L* Now a quite different approach to the problem. My idea is to associate "covariants" with the given form A(x,y,z) (J still take r = 3), as follows. P(y,z)
=:Ei
2 (A(ei,y,x» - levell, where {ei } is an arbitrary ortho-
normal basis in C n ; P is independent of the basis. Coordinate form: P(jkmn) =:Ei A
=:Eij (A(i,fj'z» 2 -
level 2.
2 R = :E ijk (A(ei,fj,gk» - level 3. The last expression is a scalar, the square of the usual HUbert(Schmidt> norm of A. 5*. At this junction it is convenient to take the step to infinite dimensional spaces. If A is a 3-linear form over a HUbert space H say taht A is "l-bounded" if P (defined as in § 4) is bounded in the usual sense. Define "2-bounded" and "3-bounded" in an analogous way. l-bouned is however trivial (;;; bounded) and so is 3-bounded (;;; HUbert-Schmidt>. So what about 2-bounded? Clearly, this is the same as tR. say that A extends to a continuous map on (H H) & H where 81 = e(projective tensor product> and ~ is the Hil'bert· tensor proauct (remark by Jonathan Arazy). Let us consider a concrete example. My problem thus now concerns the form
e.
(1)
A(x,y,z) =
~ ;(i
+ j
+ k)xiyjz k
in H = 12. Alternatively (another observation of Jonathan Arazy's) the form
J. PEETRE
218
(1')
A(£,g,h)
=
f
;P£gh
Idzl/2n
J... 2 2 2 on H = H (JJ (Hardy class). In this case H~ H = H (leT). Thus the product fgh in <1') ~y be conceived as the restriction on 'the diagonal of an t:lemen~ ofl H (X X l) and our concern is to identify the space rest H (TXT)WH (X), where rest is the map f(z,w) .. f(z,z). Now by a theorem by Horowicz and Oberlin [H4] rest H2(T X T) = AO,2(p. CI owe this reference to Eric Amar.) Here, generally spealnng, AocP = AocP(T) (oc > -1) are the weighted Bergman spaces; from our point of view th:Y are just Besov spaces: Aocp = BSP (,I? for s = -(oc + 1) Ip. By general results (cf. notably Lecture 3) p
(OC +
~)/2
and also
(AOC,l(T»* ~ Al (T) .. +oc ....
(Lipschitz space)
Combining all this information we get THEOREM.lhe form <1') is 2-bounded iff its symbol cp is in A l12 (I1. (For H = Aoc, (L) there is a similar, but more elementary result with A3oc/2+2(~')# But this is the same as the classical condition for l-boundedness ordinary boundedness); see e.g. [A3J. Thus we also get
(E:
COROLLARY. It is 2-bounded iff it is l-bounded.# This is sligthtly disappointing. It shows for instance that the ideal of 2-bounded forms probably cannot be gotten from the ideal of <1-)bounded forms and the ideal of (3-)bounded (E: Hilbert-Schmidt> forms by any reasonable interpolation process. Of course, there must exist (l-)bounded forms, which are not 2-bounded but I have no good (concrete) examples. Problem. Find one. 6*. Now we come back top the question we started out with in the first Section. How to define the analogue of the Schatten-von Neumann c\asses for multilinear forms? There is of course a good candidate for S • These are the forms ienerated by "rank one" (S-)forms. Also there is a natural choice for S : the HUbert-Schmidt ~rms. F~rthermore, the (anti-)duality coming from the inner product in S puts S in duality (at least formally) with the ideal of bounded forms, B, say. Therefore the general results on interpolation between "a space and its dual" (see Lect. 1, § 11, example), alt!!rugh rigor~slY applicable only in the reflexive case, suggest that (S ,B)1/2,2 = S , as in the classical
PARACOMMUTATORS AND MINIMAL SPACES
219
(operator) situation. Finally, this last thing suggest to define sP via real interpolation, thus: sP d~f (Sl,8)9 ' 9 = 1 - lip. (Cf. [Pl0]') With such a definition it is conceivable thatPone can extend the usual trace ideal criterion (Peller et al; cf. Lectures 3 and 4) to multilinear Hankel forms.
J. PEETRE
220
REFERENCES. [AI] Adams, R.A., "Sobolev spaces". Academic Press, New York San Fransisco, 1975. [A2] Ahlmann, M., "The trace ~eal criterion for Hankel operators on the weighted Bergman space Aa in the unit ball of Co.,. Technical report, Lund, 1984. [A3] Anderson, J.M., Clunie, J. and Pommerenke, C., "On Bloch functions and normal functions". J. reine angew. Math. 270 (1974), pp. 12-37. [A4] Arazy, J. and Fisher, S., ''The uinqueness of the Dirichlet space among Mabius-invariant Hilbert spaces". Technical report, Haifa, 1983. [A5] Arazy, J. and Fisher, S., " Some aspects of the minimal, Mabius invariant space of analytic functions on the unit disc". In: "Interpolation Spaces and Allied Topics in Analysis. Proceedings, Lund, 1983". Lecture Notes in Mathematics 1070, pp. 24-44. Springer-Verlag, Berlin - Heidelberg - New York - Tokyo, 1984. [A6] Aronszajn, N. and Gagliardo, E., "Interpolation spaces and interpolation methods". Ann. Mat. Pura Appl. 68 (1965), pp. 51-118. [Bl] Bergh, J. and Lafstram, J., "Interpolation spaces. An introduction". Grundlehren 223. Springer-Verlag, Berlin - Heidelberg - New York, 1976. [B2] Besov, O. V., "Investigation of a family of function spaces in connection with theorems of imbedding and extension". Trudy Mat. Inst. Steklov. 60 (1961), pp. 42-81 [Russian]. [B3] Besov, O.V., l1'in, V.P. and Nikol'skii, S.M., "Integral representations of functions and imbedding theorems, I-II". John Wiley, New York - Toronto - London - Sydney, 1978-79. [B4] Birman, M.Sh. and Solomyak, M.Z., "Estimates for singular numbers of integral operators. Uspehi Mat. Nauk 32:1 (1977), pp. 17-84 [Russian] • [B5] Bony, J.-M., "Cal cui symbolique et propagation des singularittls pour les tlquations aux dtlrivtles partielles nonlintlaires". Ann. Sci. Ecole Norm. Sup. 14 (1981), pp, 209-246. [B6] Brudnyf, Yu. A., "Piecewise polynomial approximation, embedding theorems and rational approximation". In: Lecture Notes in Mathematics 556, pp. 73-98. Springer-Verlag, Berlin - Heidelberg - New York, 1976. [B7] Brudnyl', Yu. A. and Kruglyak, N. Ya., "Real interpolation functors". Ookl. Akad. Nauk SSSR 250 (1981), pp. 14-17 [Russian]. [B8] Brudnyl', Yu. A. 'and Kruglyak, N. Ya., "Real interpolation functors". Yaroslavl', 1981 [Russian; English translation in preparation]. [B9] Butzer, P.L. and Berens, H., "Semi-groups of operators and approximation". Grundlehren 145. Springer-Verlag, Berlin - Heidelberg - New York, 1967. [Cll Calder6n, A.P., "Intermediate spaces and interpolation, the complex method. Studia Math. 24 (1964), PP1 113- 1gq. [C2] Calder6n, A.P. "Spaces between Land L and the theorem of Marcinkiewicz". Studia Math. 26 (1966), pp. 273-299.
PARACOMMUTATORS AND MINIMAL SPACES
221
[C3] Ceau,u, T. and Ga,par, T., "A bibliographie on interpolation of operators and applications in comutative and non-comutative harmonic analysis". [Authors' spelling!] Seminarul de Opera tori Liniari ,i Analiz~ Armonicl, special issue of 1980. Uni versi tatea din Timiioara, Sec~ia de Matematic~. [ C4] Cima, J.A., "The basic properties of Bloch functions". Internat. J. Math. Sci. 2 (1979), pp. 369-413. [C5] Coifman, R. and Rochberg, R" "Representation theorems for holomorphic functions in LP • Ast~risque 77 <1980>, pp. 11-66. [C6] Coifman, R., Rochberg, R. and Weiss, G., "Factorization theorems for Hardy spaces in several variables". Ann. Math. 103 (1976), pp. 611-635. [C7] Coifman, R., Cwikel, M., Rochberg, R., Sagher, Y. and Weiss, G., "A theory of complex interpolation for families of Banach spaces". Advances Math. 43 (1982), pp. 202-229. [C8] Cwikel, M., "K-divisibility of the K-functionl and Calderon pairs". Ark. Mat. 22 (1964), pp. 39-62. [C9] Cwikel, M. and Nilsson, P., "The coincidence of the real and the complex interpolation methods for couples of weighted Banach lattices". In: "Interpolation Spaces and Allied Topics in Analysis. Proceedings, Lund, 1983". Lecture Notes in Mathematics 1070, pp. 54-65. Springer-Verlag, Berlin - Heidelberg - New York - Tokyo, 1984. [D] Dmitriev, V.I., Krein, S.G. and Ovcinnikov, V.I., "Fundamentals of the theory of interpolation of linear operators". In: Geometry of linear spaces and operator theory, pp. 31-74. Yaroslavl' , 1977 [Russian] • [Fl] Fefferman, C., Riviere, N. and Sagher, Y., "Interpolation between HP spaces". Trans. Amer. Math. Soc. 191 (1974), pp. 75-81. [F2] Fefferman, C. and Stein, E., "HP spaces of several variables". Acta Math. 129 (1972), pp. 137-193. [F3] Flett, T.M., "Lipschitz spaces of functions on the circle and the disc". J. Math. Anal. Appl. 39 (1972), pp, 125-158. [F4] Foias, C. and Lions, J.-L., "Sur certains espaces d'interpolation". Acta Szeged 22 <1961>, pp. 262-282. [F5] Frazier, M. and Jawerth, B., "Decomposition of Besov spaces". Preprint, 1984. [Gl] Gagliardo, E., "Propriet~ di alcune classi di funzioni in piu variabili". Ricerche Mat. 7 (1958), pp. 102-137. [G2] Gagliardo, E., "A common structure in various families of functional spaces. Part II. Ricerche Mat. 12 (1963), pp. 87-107. [G3] Gagliardo, E., "Caratterizzazione construtiva di tutti gli spazi di interpolazione tra spazi di Banach". Symposia Mathematica 2 (1968), pp. 95-106. [G4] Grisvard, P., "Commutativit~ de deux foncteurs d'interpolation et applications". J. Math. Pures Appl. 45 (1966), pp. 143-290. [G5] Gustavsson, J. and Peetre, J., "Interpolation of Orlicz spaces". Studia Math. Studia Math. 60 (1977), pp. 33-59. [HI] Hahn, K. T. and Mitchell, J., "Representation of linear functionals of HP spaces over bounded symmetric domains in Co.,. J. Math. Anal. Appl. 56 (1976), pp. 379-391.
222
J. PEETRE
[H2] Hardy, G.H., "Collected papers, III". Clarendon, Oxford, 1969. [H3] Harmander, L., "Linear differential operators". Grundlehren 116. Springer-Verlag, Berlin - Gattingen - Heidelberg, 1961. [H4] Horowic:w. C. and Oberlin, D., "Restrictions of HP functions to the diagonal of U ". Indiana U. Math. J. 24 (1975), pp. 767-772. [H5] Hastings, W. H., "A Carleson measure theorem for Bergman spaces". Proc. Amer. Math. Soc. 52 (1975), pp. 237-241. "9 [Il Irodova, I.P., "On the properties of the scale of spaces B for 0 < P < 1". Dokl. Akad. Nauk SSSR 250 <1980>, pp. 273-275 [Russianf. [Jl] Janson, S., "Mean oscillation and commutators of singular integral operators". Ark. Mat. 16 (1978), pp. 263-270. [J2] Janson, S., "Generalizations of Lipschiytz spaces and application to Hardy spaces and bounded mean oscillation". Duke J. Math. 27 <1980>, pp. 959-982. [J3] Janson, S., "Minimal and maximal methods of interpolation". J. Funct. Anal. 44 (1981), pp. 50-73. [J4] Janson, S., Nilsson, P. and Peetre, J., "Notes on Wolff's note on interpolation spaces". Proc. London Math. Soc. 48 (1984), pp. 283-299. [J5] Janson, S. and Peetre, J., "Higher order commutators of singular integral operators". In: "Interpolation Spaces and Allied Topics in Analysis. Proceedings, Lund, 1983". Lecture Notes in Mathematics 1070, pp. 125-142. Springer-Verlag, Berlin - Heidelberg - New York - Tokyo, 1984. [J6] Janson, S., Peetre, J. and Semmes, S., "On the action of Hankel and Toeplitz operators on some function spaces". Technical report, Uppsala, 1984 [to appear in Duke Math. J.]. [J7] Janson, S. and Wolff, T., "Schatten classes and commutators of singular integral operators". Ark. Mat. 20 (1982), pp. 301-310. [J8] Jonsson, A. and Wallin, H., "Function spaces on subsets of R n". Math. Reports 2 (1984), pp, 1-221. [J9] Jordan, C., "R~duction d'un r~seau de formes quadratiques ou bilin~aires". J. Math. Pures Appl. 2 (1906), pp. 403-438, 3 (1907), pp. 5-51 [ reprinted in Oeuvres, t. III, pp. 269-350. Gauthiers-Villars, Paris, 1962]. [K1l Krantz, S., "Intrinsic Lipschitz classes on manifolds with applications to complex function theory and estimates for the and Bb equations". Manuscripta Math. 24 (1978), pp. 351-378. [K2] Kre'rn, S.G., Petunin, Yu.I. and Semenov, E.M., "Interpolation of linear operators". Izd. Nauka, Moscow, 1978 [Russian; English translation: A.M.S., Providence, 1982l. [L1l Lions, J.-L., "Espaces interm~diaires entre espaces hilbertiens et applications". Bull. Math. Soc. Sci. Math. Phys. R.P. Roumanie 50 (1958), pp. 419-432. [L2] Lions, J.-L., "Th~oremes de traces et d'interpolation, I-V". Ann. Scuola Norm. Sup. Pisa 13 (1959), pp. 389-403, 14 (1960), pp. 317-331; J. Math. Pures Appl. 42 (1963), pp. 195-203; Math. Ann. 151 (1963), pp. 42-56; An. Acad. Brasiliera Cien. 35 (1963), pp. 1-10. [L3] Lions, J.-L., "Equations diff~rentielles o~rationnelles et problemes mixtes". Grundlehren 111. Springer-Verlag, Berlin Gattingen - Heidelberg, 1961.
a
PARACOMMUTATORS AND MINIMAL SPACES
223
[L4] Lions, J.-L. and Magenes, E., "Problemes aux limites non homogEmes et applications, I". Dunod, Paris, 1968. [L5] Lions, J.-L. and Peetre, J., "Sur une classes d'espaces d'interpolation". Publ. Math. Inst. Hautes Etudes Sci. 19 (1964), pp. 5-68. [M] Mityagin, B., "An interpolation theorem for modular spaces". Mat Sbornik 66 (1965), pp. 473-482 [Russian; English translation in: "Interpolation Spaces and Allied Topics in Analysis. Proceedings, Lund, 1983". Lecture Notes in Mathematics 1070, pp. 10-23. Springer-Verlag, Berlin - Heidelberg - New York - Tokyo, 1984]. [Nl] Nikol'skii, S.M., "Approximation of functions of several variables and imbedding theorems". Grundlehren 205. Springer-Verlag, Berlin - Heidelberg - New York, 1975. [N2] Nilsson, P., "Reiteration theorems for real interpolation and approximation spaces". Ann. Mat. Pura Appl. 132 (1982), pp. 291-330. [N3] Nilsson, P., "Interpolation of Calderon pairs". Ann. Mat. Pura Appl. 134 (1983), pp. 201-232. [01] Oklander, E. T., A interpolators and the theorem of Marcinkiewicz". Bull. Amer. MathP§oc. 72 (1966), pp. 49-53. [02] Ovchinnikov, V.I., "Interpolation theorems resulting from Grothendieck's inequality". Funkcional. Anal. i Prilozen. 10 (1976), pp. 45-54. [03] Ovcinnikov, V.I., "The method of orbits in interpolation theory". Math. Reports 1 (1984), pp. 349-515. [Pll Peetre, J., "Espaces d'interpolation et th~oreme de Soboleff". Ann. Inst. Fourier 16 (1966), pp. 279-317. [P2] Peetre, J., "N~w thoughts on Besov spaces". Duke Univ. Math. Series 1. Durham, 1976. [P3] Peetre, J., "Invariant function spaces connected with the holomorphic discrete series". In: Butzer, P.L. et al (eds.), "Anniversary volume on Approximation Theory and Functional Analysis", pp. 119-134. Birkhluser, Basel, Boston, 1984. [Also available in the Lund technical report series.] [P4] Peetre, J., "Recent progress in real interpolation". In: "Methods of Functional Analysis and Theory of Elliptic Equations Proceedings of the International Meeting dedicated to the memory of professor Carlo Miranda, Naples, September 13-16, 1982", pp. 231-263. Liguori, Naples, 1983. [P5] Peetre, J., "The theory of interpolation - its origin, prospects for the future". In: "Interpolation Spaces and Allied Topics in Analysis. Proceedings, Lund, 1983". Lecture Notes in Mathematics 1070, pp. 1-9. Springer-Verlag, Berlin - Heidelberg - New York - Tokyo, 1984. [P6] Peller, V.V., "Hankel operators of class S and applications (rational approximation, Gaussian processes, the m~orant problem for operators)". Mat. Sb. 113 <1980>, pp. 538-581 [Russian]. [P7] Peller, V. V., "Hankel operators of the Schatten-von Neumann class S , 0 < P < 1. LOMI preprints E-6-82, Leningrad, 1982. [P8f Peller, V. V., "Estimates of functions of power bounded operators in Hilbert space". J. Operator Theory 7 (1982), pp, 341-372. [P9] Persson, L.-E., "Description of some interpolation spaces in off-diagonal cases". In: "Interpolation Spaces and Allied Topics in Analysis. Proceedings, Lund, 1983". Lecture Notes in Mathematics 1070, pp. 213-231. Springer-Verlag, Berlin - Heidelberg - New York - Tokyo, 1984.
"
224
J. PEETRE
[P10] Pietsch, A., "Ideals of multilinear functionals". Preprint, Jena, 1983. 1 [Rll Rivi~re, N.M. and Sagher, Y., "Interpolation between L· and H , the real method". J. Funct. Anal. 14 (1973), pp. 401-409. [R2] Rochberg, R., "Trace ideal criteria for Hankel operators and commutators". Indiana Univ. Math. J. 31 (1982), pp. 913-925. [R3] Russo, A., "On the Hausdorff-Young theorem for integral operators". Pac. J. Math. 16 (1977), pp. 241-253. [R4] Rubel, L.A. and Timoney, R.M •• , "An extremal property of the Bloch space". Proc. Amer. Math. Soc. 75 (1979), pp. 45-49. [R5] Rochberg, R. and Weiss, G., "Derivatives of analytic families of Banach spaces". Ann. Math. 118 (1983), pp. 315-347. [Sll Semmes, S., "Trace ideal criterion for Hankel operators, 0 < p < 1". Integral Equations Operator Theory 7 (1984), pp. 241-281. [S2] Shapiro, H.S., "A Tauberian theorem related to approximation theory". Acta Math. 120 (1968), pp. 279-292. [S3] Shields, A., "The analogue of the Fejllr-Riesz theorem for the Dirichlet space". In: Beckner, W. et al (eds.), "Conference in Harmonic Analysis in honor of Antoni Zygmund", vol. II, pp. 810-820. Wadsworth, Belmont, 1983. [S4] Stein, E.M., "Singular integrals and differentiability properties of functions". Princeton University Press, Princeton, 1970. [S5] Stein, E.M., "Singular integrals and estimates for the CauchyRiemann equations". Bull. Amer. Math. Soc. 79 (1973), pp. 440-445. [56] Stein, E.M. and Weiss, G., "Interpolation of operators with change of measure". Trans. Amer. Math. Soc. 87 (1958), pp. 159-172. [S7] Strichartz, R., "Para-differential operators - another step forward for the method of Fourier". Notices Amer. Math. Soc. 29 (1982), pp. 440-445. [Tll Taibleson, M., "On the theory of Lipschitz spaces of distributions of Euclidean n-space. I-II". J. Math. Mech. 13 (1964), pp. 821-839. [T2] Triebel, H., "Interpolation theory. Functions spaces. Differential operators". VEB, Berlin, 1978. [T3] Triebel, H., "Theory of function spaces". Akademische Verlagsgesellschaft Geest & Portig, Leipzig, 1983. [T4] Timotin, D., "A note on C estimates for certain kernels". Preprint, Bucuresti, 1984. p [U] Uchiyama, A., "On the compactness of operators of Hankel type". Tohoku Math. J. 30 (1978), pp. 163-171. [W] Wolff, T., "A note on interpolation spaces". In: "Proc. Conf. Harmonic Analysis, Univ. of Minnesota, Minneapolis". Lecture Notes in Mathematics 908, pp. 199-204. Springer-Verlag, Berlin - Heidelberg New York, 1982. [Z] Zygmund, A., "Trigohometrical series I-II". Cambridge University Press, Cambridge, 1958.
DECOMPOSITION THEOREMS FOR BERGMAN SPACES AND THEIR APPLICATIONS
Richard Rochberg
1
Washington University, St. Louis, Missouri
We sketch a decomposition theorem for functions in Bergman spaces and give applications. Bergman spaces consist of holomorphic functions which are integrable with respect to area measure. Such functions can be decomposed as norm convergent sums of rational functions. We present that theorem, its variations and applications. We obtain inclusion relations between the various Bergman spaces, Sobolev style embedding theorems and results on interpolation of values by functions in the Bergman spaces. Decomposition theorems are also discussed for certain Besov spaces and for BMO. These theorems have consequences in the theory of rational approximation. The decompositions are also used to study linear operators which have a "symbol" in a Besov space. We give boundedness and Schatten ideal results for commutators, Hankel operators, and products of Hankel operators. The Hankel operator results can be used to obtain converses of the earlier results on rational approximation. INTRODUCTI~
The Bergman spaces, which we will denote functions
AP~, are spaces of
f which are holomorphic in the upper half plane
which satisfy 225 S. C. Power fed.), Operators and Function Theory, 225-277. © 1985 by D. Reidel Publishing Company.
U and
R.ROCHBERG
226
P = lIfll po,
Iu
If (x+iy) I p yo,dxdy
The range of interest is
O
-1<0,.
The decomposition
theorems we present show that in these and related spaces it is possible to find elements which behave similarly to the standard basis elements in the sequence spaces t P • For each p and a, we will find elements and only i f numbers
f
[Ai}
such that a function
can be written in the form in the sequence space t P •
is in
f
f=L: A.b. 1
1
APo,
if
with the
In Section 2 we give
a precise statement of this and related results and an outline of the proof.
Roughly the proof consists of first representing
f
with an integral reproducing formula and then approximating the integral by a Riemann sum. Once we know how to build up and break apart functions in various spaces, many results about linear maps into and out of the spaces become relatively simple.
In other cases the decomposition
theorem lets us break up operators into easy to understand bits. These are the main themes of the applications in Section 3. In Section 3A we give the inclusion theorems between the Bergman spaces and also the Sobolev type embedding theorems. basic idea is that each individual building block
bi
The
sits in a
whole family of spaces and also behaves well under differentiation and integration.
The
b.
1
are rational functions and we also give
results about rational approximation. In Section 3B we discuss Hankel operators. acting on the Hardy space and projection.
H2
These are operators
which are built out of multiplication
The decomposition theorems are used two ways.
First, the function which does the multiplication can be decomposed. This splits the operator into small (i.e. finite rank) pieces. Second, the decomposition theorem applied to
H2
produces a set
of functions which are a good substitute for an orthonormal basis which diagonalized Hankel operators.
Using these tools we give
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
227
conditions for a Hankel operator to be in a Schatten ideal. The decomposition theorems let us build functions with desired properties by selecting the
A.
1
properly.
In Section 3C we use
that idea to construct functions in the Bergman spaces which take assigned values at specified points. The results in the first three parts of Section 3 use norm estimates for the
A.. 1
The results in Section 3D use more refined
estimates on these coefficients.
We describe how to use the
decomposition of function of bounded mean oscillation to obtain boundedness results for first and second order commutators.
We
describe refined estimates of singular numbers of Hankel operators and products of Hankel operators.
We also present results on
interpolation of values with pointwise control of the interpolation function. There are some ways in which these notes are a misleading introduction to decomposition theorems for Bergman spaces.
First
of all, the decomposition theorems involve a fair amount of technical machinery.
One of the rewards for this technical investment
is that the theory extends with very little change to certain spaces of harmonic functions and of holomorphic functions of several complex variables.
However, we have not included those extensions.
Second, our main theme is the applicability of the decomposition theorems of [CR] and [RS].
Little notice is given to the fact that
many of the results presented here were first obtained (or can also be obtained) by other methods, with alternative insights and, in some cases, more direct proofs.
In particular, we will not present
the rich theory of Hankel operators as developed by Peller.
Finally,
there are other approaches to these decomposition theorems, in particular that of Luecking and those based on A. P. Calderon's reproducing formula, which we do not present here at all.
For
these extensions and alternative approaches we have done little more than provide recent biographical references.
228
R.ROCHBERG
We will use the letter "c" as a generic constant and will set to one various constants in Fourier transform formulas. THE DECOMPOSITION THEOREM
A.
Background When studying a space of functions, it is very useful to be
able to write every function in the space as a linear combination of functions which are, in some appropriate sense, elementary.
The
prototype of such decompositions is the natural expansion of elements tP
in the sequence space
in terms of the standard basis elements.
A more subtle example is given by the Fourier representation of functions in L2 of the unit circle: f is in L2 if and only if
Furthermore
The decomposition of
t P was in terms of the supports of functions.
This decomposition of functions in
L2
is in terms of frequencies.
The decomposition which we will give for holomoprhic functions mixes the two points of view.
We will try to break a function into
localized pieces; however, since holomoprhic functions are never compactly supported, complications arise.
The terms we obtain can
be thought of as indexed by both location and frequency. The atomic decomposition of the Hardy space
HI
was the
motivating example for the decompositions we will consider for the Bergman spaces.
Here is the decomposition of
HI
(in a version
which does allow compactly supported functions as building blocks). For a function by
f(x,y)
half-plane
f(x)
defined on the real axis
the harmonic (Poisson) extension of U
is defined by
f
R
to the upper
The non-tangential maximal function of Mf(x)=sup[\f(x,y)\;\x-y\<\y\}.
we denote f,
A function
Mf, f
is
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
in the Hardy space == IIMfll l ' L
HI
if
is finite.
consider distributions when An atom is a function I,
an interval
[ail
p,
Ll. Norm HI by IIflll P H is the space of functions
(Actually, it is necessary to p< 1 .)
a(x)
defined on
If
f
so that
R,
supported on
having mean value zero, and normalized by
The atomic decomposition of
IIrl .
Theorem:
is in
Similarly, for all
for which
~
Mf
229
is in f
HI
HI
is the following
then there are numbers
[Ai}
and atoms
can be written (2. 1)
f
with
lal
~ IAil -;; cllfll l
by (2. 1) is in Thus norm on
HI HI
HI
Conversely, i f and satisfies
~ IA.I < 1
CD
then
f
defined
I\fl\l-;; c~IAil.
consists exactly of sums of the form (2.1) and the is equivalent to the "atomic norm": f
given by (2.1)} •
Unfortunately we must refer to [CW] or [G] for a description of the (very close) relation between the space we just described and the space of holomorphic functions studied by Hardy.
Roughly,
we have just described the boundary values of the functions studied by Hardy.
[CW].
Many applications of this decomposition are given in
As to the proof, this result is almost a reformulation of
C. Fefferman's celebrated result that the dual of of functions of bounded mean oscillation.
HI
is the space
We will say more about
that space later. ~xercise
for the reader:
if the atoms are not required to
have mean zero then the analogous (but much easier) result is an atomic decomposition of the Lebesgue space This result is a model for our results.
Ll.) There are several
R.ROCHBERG
230
things to note about the decomposition.
Each function in the space
is represented as a sum of localized building blocks. block has a characteristic location (the center of characteristic scale (the length of
I).
bits has some built in cancellation.
Each building
I ) and a
Each of the localized
Finally, with minor modifi-
cation in the definition of atoms, and an appropriate change in normalization, a similar decomposition holds for HP , p < I . The main change is that the coefficients must satisfy B.
The Bergman Spaces on
1.
fractional maps of
U
d(z,w),
The hyperbolic metric on
U is
U which is invariant under the linear
to itself and which agrees infinitesimally
with the Euclidean metric at for
=.
U
Hyperbolic geometry.
the distance function on
~ 1AilP <
z = i.
Exact formulas can be given
the hyperbolic distance between
z
and
w,
but
the following approximation gives the idea more effectively:
(Here and throughout we will write
z = x + iy,
w = u + iv,
and
,=g+i'Tl .) We will call a sequence of points every point of zi'
in
U d-dense if
U is within (hyperbolic) distance
We will call a sequence
sequence are within distance sequence a
[zi}
d
of some
s-separated if no two points of the s
of each other.
d-Iattice if it is both
Sd-dense and
We will call a dIS-separated.
For example, the set of points
is a
I-lattice. ~xercise:
every set of points that is
d-Iattice. ) '!he decomposition theorem. 2.
d-dense contains a
The basic decomposition
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
231
A~
theorem states that functions in the various
can be repre-
sented a~ sums of normalized translates and dilates of with coefficients in Theorem 2.2: d
-r,p.
Suppose
(2 + a.)max (1, lip). and any
(z + i)-n
0< p <
co,
-1 < a.
There is a positive
d-lattice
Pick d
o there is a
[zi}
a
with
a
>
such that for any c=c(a.,p,a,d)
so
that (a)
APa.
is in
f
if
then
a-(2+a.)/p
Yi
f (z)
I: Ai
(2.3)
- a (z - z. ) 1.
with
I: IAilP ~ cllfll~a. • (b)
Conversely, i f
converges in a function
APa. f
in
I: IAil p
norm and uniformly on compact subsets of APa.
Apa.
U
to
IIfll:a. ~ c I: IAilP
which has
(Note tha t we are writing Thus
is finite then the sum in (2.3)
z. =x.+iy .• ) 1.
1.
1.
consists of sums of the form (2.3) and the norm
is equivalent to the decomposition norm: IIfll P.... inf[ I: II.. .I P ; 1.
f
is given by (2.3)}
Although the representation in (2.3) is not unique, the proof will obtain A.
1.
function
for
which depend linearly and continuously on the
f
There are many similarities to the decomposition described HP , P ~ 1. As a function on the real axis each building
block has a characteristic location (that is, dispersion (given by
y i).
x. ) and a scale of 1.
The form of the individual terms
insures that they have mean zero (and higher moments zero if is small).
p
The cancellation is built into the form of the indi-
Vidual terms.
The same building blocks, with renormalization,
work for a range of
p.
232
R.ROCHBERG
3.
Proof discussion.
We will discuss the main ingredients
of the proof and how they fit together.
A full proof is given in
[CR] and a proof of a special case which contains many of the main
ideas is given in [P02]. The power of
in
is exactly what
is needed to insure that such terms have an independent of atelyif
z..
norm which is
Part (b) of the theorem now follows irmnedi-
~
p
AP~
(For
p< 1
IIfllp~
is not a norm but a quasi-norm.
However it does satisfy
Thus for
p< 1
it is often relatively easy to add individual
estimates. ) We now discuss part (a) in the case
p = l,
~=
O.
The
starting point for the proof is the reproducing formula
f (z)
Here a
=2
, =
~
+ iT] and
dV
is the Euclidean volume element.
this formula is a consequence of Cauchy's formula.
If
Some
other cases can be obtained from that one by integration by parts. The general formula can be obtained using the Fourier transform or by using the theory of Hilbert spaces with reproducing kernels (for the Hilbert spaces
A2~).
A proof can be found in [K].
The idea now is to approximate this integral by a Riemann as the union of sets B.~ so that zi is in B. and the (hyperbolic) diameter of each Bi is of order of ~ 2 magnitude d • The Euclidean volume of Bi is roughly Yi and sum.
hence
Partition
U
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
233
2 ,.., c 1: f (z .) y. 1
1
This is approximately the required decomposition with 2
cf(z.)y .• 1
Ai
Notice that the sum
1
is a Riemann sum for
IIfli
and hence can be contro1ed by
Ilfli.
To make this idea into a proof, it is enough to show that the approximating sum is close to
f
is the sense that for some
c
\If - approximation\l < c < 1 • The constant
c
which shows up in the proof of this estimate
involves a factor of for small
d.
d.
Hence the required estimate is possible
Once that estimate is obtained, an iteration of the
approximation scheme will give the required decomposition.
A and use the Neumann
is, write the approximation operator as series:
-1
A
= (I-(I-A»
-1
=1: (I-A)
(That
n
.)
The norm estimates are based on estimating the difference ]
a - 2
- a (z - , )
for
z
in the set
- f (z. ) 1
- a (z - z. ) 1
B •• 1
The important fact about these estimates
is that the natural modulus of continuity estimates for f and a - 2 --a are in terms of the hyperbolic geometry. Thus for Yi (z - zi) the difference can be controlled in a useful way by the integral of
If I
over a large hyperbolic ball containing
Bi
R.ROCHBERG
234
These ideas,
are enough to complete the proof for
the
Ifl P
together with the fact that p
~
is subharmonic,
1 •
When
p> 1
[A.}
and the norm estimates on the individual terms in (2.3)
1.
a new complication shows up.
The summability of
are not enough to insure that the sum (2.3) converges.
One way to
establish the convergence is to dominate the sum by an integral. That is If
Lemma 2.4:
p> l ,
a> 2,
then the map of
f
to
Tf
given
by
Tf (z)
is a bounded map of
L (U,dxdy)
to itself.
This lemma is an immediate consequence of the following two lemmas. Lemma 2. 5:
If
a> 2
Lemma 2.6:
Suppose
q=p/(p-l)
+ ex,
then
1< P<
Suppose
00
and let
Q(z,w)
If there is a positive function < c gq (z) f
to
Tf
map of
q
be the conjugate index,
is a positive function on g such that
Ju Q(z,w)gq(w)dV(w)
Ju Q(z,w)gP(z)dV(z)
and
LP(U,dV)
U XU.
the map of a bounded
to itself.
Lemma 2.5 can be established by Plancherel's Theorem.
(In
fact that gives the lemma with equality for appropriate c.) Lemma 2.6, which goes back to I.
Schur, rests on Holder's inequality.
To obtain Lemma 2.4, use the boundedness criterion of Lemma 2.6 with the test function insnrps tha t
e
g(z)=ye
CCln be found.
for appropriate
e·
Lemma 2.5
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
235
To use Lemma 2.4 note that the right hand side of (2.3) is dominated by
TH with
-
H(z) -
~
1A. 1 y.-2/p ~
~
XB (z).
.
(Again, here
~
we use the fact that the natural modulus of continuity estimates are in terms of the hyperbolic geometry.)
Direct computation shows
Lemma 2.4 is also used to pass from pOintwise estimates for
f - Af
p>1 .
ing par t a for
The proof of the theorem for pattern.
to norm estimates when prov-
cx," 0
follows the same general
The local modulus of continuity estimates are almost
unchanged because powers of
yare roughly constant on each
Bi
Lemmas 2.5 and 2.6 are sufficient to establish an analog of Lemma 2.4 for the space LP (U,ycx, dxdy) • C.
Variations 1.
Other'domains.
proof of Theorem 2.2:
There were two basic tools used in the first, a reproducing formula which gives a
bounded operator even after absolute values are brought inside the integral sign; second, an underlying geometry with respect to which the kernel of the reproducing formula (and hence also the functions being reproduced) satisfy convenient modulus of continuity estimates. To some extent both of these tools are available for spaces of holomorphic functions defined on symmetric domains in Cn (such domains have a transitive group of conformal automorphisms) and which are integrable with respect to (possibly weighted)Euclidean vohnne.
(These are the general Bergman spaces.)
The appropriate
reproducing formulas involve integration against the Bergman kernel function (or its powers).
Fourier transform techniques can be
used to obtain an analog of Lemma 2.5.
The analog of Lemma 2.6
is true for any measure space and hence a version of Lemma 2.4 can be obtained.
The transitive group of automorphisms can be used
to reduce local questions about the modulus of continuity to a
236
R. ROCHBERG
fixed base point.
Local questions at a fixed base point can then
be analyzed directly. For instance, here is the version of Theorem 2.2 for functions defined on the unit disk in in
en.
or, more generaly, the unit ball
The Bergman kernel for the ball is B(z,w)=c(1-z.;)-(n+I).
We consider spaces
Ilflli
C
I
pr
Iz
A'
of holomorphic functions for which
pr
J r< I
If(z)IPB:z,z)-r dV(z)
(The Bergman kernel for
U
- -2 c (z - w) ,
is
hence
r
in
Apr
). A lattice is defined as before pr in terms of the invariant distance and Theorem 2.2 is true as corresponds to
r/2
A'
in
stated with the basic decomposition formula (2.3) replaced by B (z , z.)
f (z)
r:
f... • 1
B(z., z.)b - (l+r) /p 1
with
b
1
1
b>(I+r)max(1, lip). In this formulation the theorem holds for other domains also,
see [CR] (but also [CR2 D.
The decomposition given in (2.7)
involves powers of the Bergman kernel evaluated at points of a lattice.
(However, other analogs of these results suggest that
the basic building blocks can more profitably be thought of as derivatives (or certain directional derivatives) of the Bergman kernel, rather than as powers of the kernel.) The Hardy space H2. For 0 < P < co, the (holomorphic) Hardy space HP • is that space of functions which are holomorphic 2.
in
U and for which
When convenient, these functions will be identified with their
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
boundary values on
~.
237
(Unfortunately, the space we called
earlier is not the space we just defined with
p=1
HI
Precisely,
the space defined before is the complexification of the space of real parts of the boundary values of the functions in the space just defined, see [CWJ).
A function is in is supported on AP~
H2
(0,00)
if and only if its Fourier transform 2
and is in
L «O,oo),dt).
need not have pointwise boundary values on
however, have boundary distributions.
Functions in the
R.
They do, A function is in A20 if
and only if its boundary distribution has Fourier transform sup2 -1 ported on (0,00) and in L «0,00), t dt). Thus integration of order one-half (which, by definition, means dividing the Fourier transform by t 1l2 ) is an isomorphism of to H2. We can
iO
use this fact to obtain a decomposition theorem for functions in H2 . Theorem 2.8: isa
Fix
c=c(b,d) (a)
if
b > 3/2.
d
If
so that for any f
f (z) = I: A.
1
is in
H2
is sufficiently small then there d-lattice
[Zi} ,
then
(2. 9)
b (z-z. ) 1
is finite then the sum in
Conversely, if
(b)
(2. 9) converges in norm and uniformly on compact sets to a function f
in
H2
which satisfies
Ilfll22 -; cI: 1Ail2 H
This is a direct consequence of Theorem 2.2, of
A20
to
H2,
~f
the relation
and of a Fourier transform calculation which
shows that half-order differentiation and integration have the -a expected effect on the functions (z - z.) • The requirement 1
238
R.ROCHBERG
b> 3/2 for
is an artifact of the proof.
b> I
as is shown by Luecking [L3] using a different proof.
The relation between the when
p=2
for other 3.
The result is actually true
A20
and
HP
is this simple only
In fact, the tempting analog of the previous theorem p
is false [RS].
Bloch, Besov, and BMO.
In this section we discuss the
the decomposition theorems for certain (diagonal) Besov spaces, for the Bloch space, and the space mean oscillation.
BMO
of functions of bounded
The Besov spaces we consider are obtained from
Bergman spaces by integration and hence the decomposition theorems will follow from Theorem 2.2 for
O
The Bloch space will
be seen as a limiting case of these results for it requires a separate proof.)
The space
BMO
Bloch space and the Besov spaces we consider.
p = ex:>
(although
fits between the It also can be
regarded as a limiting case of the Besov spaces but in a more de lica te way. We have been considering (interchangeably) holomorphic functions defined on
R R
U and their boundary distributions defined on
There are analogous results for spaces of distributions on which are not of holomorphic type.
The difference in the decom-
position theorems is that two lattices must be used •.. one in
U
and one in the lower half plane
L
will involve the derivatives in
U of the holomorphic extension
to
The definitions we will give
U of distributions of holomorphic type.
The definition of
the corresponding general spaces of distributions uses the harmonic extension of the distribution to For
O
select
to be those functions
f
m>l/p
U and partial derivatives. and define the Besov space
which are holomorphic in
U
BP
for which
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
239
It is a consequence of the general theory of Besov spaces, or of Theorem 2.4, that the space so defined does not depend on For any
m> I
m.
the space of Bloch functions is defined to be
the space of
f
for which
sup U
Again, the definition can be shown to be independent of different
m
To define measure.
m
with
giving equivalent norms. BMO
A measure
we first recall the definition of a Carleson u
U
on
is said to be a Carleson measure
if
sup { I a xElR a>O
I~!((x-a,
x+a)x(O,a))}<
ex>
A holomorphic function is said to be in the space ylf' (z)!2 dxdy llfl\BMo
by
is a Carleson measure.
BMO
We define the
if
BMO
norm,
Although this definition
I\fl\BMO
will be convenient later,
it is not the traditional defintion.
Traditionally a function
f
is said to be in
BMO
of the line
if
sup xER a>O
I (inf 2a c
x+a
J
!f(t) - cldt) <
ex>
x-a
With this definition it is relatively easy to see that ,the dual of the space given earlier.
HI
The fact that the two definitions of
for holornorphic functions is not so obvious. various descriptions of
BMO
is
for which the atomic decomposition was
BMO
see [G],
It is easy to check that i f
p< r
[Ko]. then
BMO
agree
For more about the
240
R.ROCHBERG
BP
C
Br c
BMO
Bloch.
C
Here are the decomposition theorems for these function spaces. Theorem 2.10: (b> 1
Suppose
O
for the case of Bloch or
small then there is a
Pick BMO.)
c=c(d,b,p)
b>max(O, 1- lip) • If
d
is sufficiently
so that for any
d-lattice
{zi} (a)
if
f (z)
f
is in
l:"A.
1
B
-
(z - z" )
then
b
1
1"- .I P -s;
l:
with .
1
cUfllP
BP
-
.
f
If
is in the Bloch space then sup£! A)} ;; cllfllBloch
can be represented as in (2.11) with f
is in
BMO
then
f
f If
can be represented as in (2. 11) and the
numbers satisfy a quadratic Carleson measure condition
(Here
6
is the point mass at
z"1
(b)
Conversely, i f
converges in f
in
BP
BP
l: ILIP 1
is finite then the sum (2.11)
norm and uniformly on compact sets to a function
which satisfies
then the sum (2. 11) fies
z1".)
llf ll : p ;; cL:I"-i IP .
c~nverges
IIfllBloch;; c sup £lAiD.
If
sup£1AJ}
to a function in Bloch which satisIn this case the convergence is
weak* convergence with the Bloch space realized as the dual space 10 2 of A • If L: A" y" 6 is a Carleson measure then the sum in 1 1 zi
(2.11) converges to a function in
BMO
which satisfies
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
IIflliMO -;; cllL: 1Ai\2Yi 6z
.lI cM
·
241
In this case the convergence is
~
in the weak* topology with
BMO
regarded as the dual of
In these last two cases, Bloch and need not converge pointwise.
BMO
HI.
the series (2.11)
However, if an appropriate constant
is subtracted from each term then it is possible to obtain convergence which is uniform on compact subsets of
U.
HI
as the dual of
have mean zero.
Hence if we regard
then the elements of constants. space.
BMO
(Functions in HI
BMO
are only well defined modulo additive A similar observation holds for AlO and the Bloch
This is why we only get pointwise convergence only after
subtracting appropriate constants.) Note that for any
b> 0 ,
have maximum modulus of
1
As we mentioned, the
on BP
b --b y. (z-z.) ,
the individual terms
~
~
U. case of the theorem follows from
Theorem 2.2 by term by term integration.
To obtain the Bloch
version of the theorem, first prove an analog of Theorem 2.2 for holomoprhic functions
g
which have
y\g(z)1
(The proof extends without difficulty.)
bounded in
Since such
g
U
are exactly
the derivatives of Bloch functions, the required representation for Bloch functions now also follows from term by term integration. (The proof in [eR] is more awkward. ) The decomposition of a function in starting with the derivative of
f,
BMO
is also obtained by
following the pattern of
the proof of Theorem 2.2 and then integrating the resulting sum term by term.
The proof that such a sum is in
BMO
is obtained
by a direct estimate of the Carleson measure norm of
y 1f'I2 dxdy .
This is given in detail in [RS] where there is also a description of how the individual coefficients
Ai
in the decomposition can
be estimated in ways that emphasize the behavior of
f
near
z
There is natural control at the first step of the approximation
242
R.ROCHBERG
process.
The issue is keeping control as the approximation pro-
cess is iterated.
That type of local control is useful in the
applications given in the last section. 4.
Other spaces.
We have been concentrating on decomposi-
tions of Bergman spaces of holomorphic functions.
Similar tech-
niques can be used to give decomposition theorems for spaces of harmonic functions, including mixed norm spaces with norms such as
IIfll
co
co
o
-co
(J (J
I
\f(x,y)\rdx)syi3dy)rs
That theory, including the application to the description of certain mean oscillation spaces, is given in full by Ricci and Taibleson [RT] for R and by Bui for Rn [B]. Similar results for various mean oscillation, Beurling, and Hardy spaces have been given by Chao, Gilbert, and Tomas [CGT], [CGT2], and by Merryfield [M]. decomposition of tion of 5.
HI
Their results include the atomic
(based on the square function characteriza-
HI.) Other proofs.
Our analysis was based on reproducing
formulas which used the Bergman kernel function.
There is another
type of reproducing formula which was introduced by A.P. Calderon and which has been very useful in obtaining decomposition theorems for various function spaces.
In particular it is at the heart of
Uchiyama's decomposition for
BMO
the atomic decomposition of Chang and R. Fefferman [CF].
HI
[U] and
~
Wilson's proof of
[W] as well as earlier work of
Applications of that point of view
to the Besov spaces are given by M. Frazier and B. Jawerth [FJ]. Related work has been done by G. Cohen [C]. Luecking has developed techniques for analysis of functions in the Bergman spaces
whi~h
yield
theorems we described [LI,2,3].
many of the decomposition
Although his approach is less
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
243
constructive (since he uses a duality argument from functional analysis at a critical stage) it gives a clearer insight into the role of the local geometry and the dual nature of the conditions that a sequence be dense and that it be separated.
He also gets
decomposition theorems for weighted space---a context in which conformal automorphisms are less useful. 6.
Atomic and molecular decompositions of other spaces.
Decompositions of the type we have been describing were first used systematically for the Hardy spaces
[CW].
A good way to
find out about recent activity in that area would be to scan the conference proceedings [BCFJ]. Various other function spaces have also had their members dissected in similar ways recently.
A decomposition theorem for
a Sobolev space, together with various applications is given by Jodiet in [J].
The spaces of functions made of blocks (not the
Bloch space) was first introduced by Taibleson and Weiss via an atomic decomposition.
For more on these spaces see [TW] and [So].
The tent spaces of Coifman, Meyer, and Stein are closely related to the Hardy spaces and have similar decompositions.
(See
[CMSI,2]. ) APPLICATIONS Many applications of atomic and molecular decompositions use the fact that the individual building blocks satisfy size, localization and cancellation estimates.
Although some of our applica-
tions are of this sort,most of them exploit the explicit form of the terms in Theorem 2.2. Some of the results we describe could be obtained by working directly with the reproducing formula. proofs using quite different ideas.
Others have alternative
Our main theme here is that,
once the decomposition theorem is available, many results follow quiet naturally and easily.
To emphasize the simplicity of the
method we present some results in less than full generality.
R.ROCHBERG
244
However, a virtue of the method is that once the proof is given for the simple case, the more general proof is often identical. (For instance, many of the proofs extend directly to Bergman spaces of functions of several complex variables. ) A.
Inclusion Theorems, Rational Approximation and Interpolation For fixed
p
p
and variable
or for fixed
~
and variable
~
there are no inclusion relations between these spaces.
However,
if both parameters vary at once then there is a family of inclusions.
This is an immediate consequence of the form of the
individual terms in (2.3), the fact that for fixed
p
Theorem 2.2 can be used with any sufficiently large inclusion relations between the sequence spaces t P Theorem 3.1:
~
and
a,
and the
That is,
Suppose
o< P < pi,
(2
+ ~) / p =
(2
+ ~I
)
/
p,
(3.2)
•
Then
the inclusion is continuous, and if
Al~1 in
pi = 1
of the convex hull of the unit ball of
then the closure in
AP~
contains a ball
Al~'
Thus, informally, taining
A~.
Corollary 3.3:
Al~I
is the smallest Banach space con-
Precisely The spaces
AP~
and
Al (2 +~) /p
have the same
duals. We have not defined Bergman spaces with
~
= -1.
With
appropriate normalization that limiting case correspond to having norms defined by intergation along the real axis.
This suggests
that we might interpret the Hardy space (which, for holomorphic functions can be defined in terms of the LP(R,dx) norm of their boundary values) as the spaces AP - 1. Although not all the
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
245
conjectures suggested by this analogy are correct, in this case we have Theorem 3.4:
O
If
and HP
is a continuous inclusion of
lip = (2+o.,')/p' into
A P ex,
The closure of
the convex hull of the image of the unit ball of A1, lip - 2
and
contains a ball in
AI, lip - 2
then there
, ,
AI, lip - 2.
HP
in
In particular
HP
have the same dua 1 space.
Theorem 3.1 is a direct consequence of Theorem 2.2. In Theorem 3.4 the continuous inclusion of HP in the Bergman space is a theorem of Hardy and Littlewood. claim
To demonstrate the second
it suffices, by Theorem 2.2,to show that the functions
are uniformly (with respect to compute the
LP(R,dx)
, ) in
HP •
To check this we
norm of the boundary value.
That norm is
the integral of a constant times the Poisson kernel and is independent of
,.
The identification of the dual spaces of first given by Duren, Romberg, and Shields.
HP
for
P< 1
was
They followed this
path of first identifying that Bergman space which was the smallest containing Banach space of HP For more details and references see [CR]. It is straightforward to find the effect of differentiation and integration on functions in the Bergman spaces.
We define
integration of arbitrary order using the Fourier transform. any complex number the
APex,
Here
fA
w with
Re (w) > 0
we define the integral of
and function f
of order
w,
f
For
in any of IWf,
denotes the Fourier transform of the distributional
by
R.ROCHBERG
246
boundary values
lim f(x+iy).
These operators have the expected
y-O
r
action on basic building blocks.
w
-
«z - ' )
-a
-
) = c (z - , ) DW
We define the general differentiation operators
by
-a +w DW
•
= r- w
Here is the extension of Theorem 3.4. Theorem 3.5:
Suppose
p,p'
satisfy
a,a'
w
Re (w) ~ 0
is a complex number,
and
(2+a)/p- (2+a')/p=Re(w) ,
rW
then
is a continuous map of
into
Apa
I
aP a
This is the extension to the Bergman spaces of the classical Sobolev embedding theorems. case
(Note that in the formal limiting
we obtain the indices for the Sobolev embedding
a = -1
theorems for
HP~).)
The theorem also shows that integration of purely imaginary order is an isomorphism of each Apa to itself. The theorem is not quite a consequence of Theorem 2.2.
The
general integration operators certainly map the sum given by (2.3) to another sum of the same type, but involving the complex exponent a - w.
Writing
(z _ '" ) -a
Note that of
(;
w = u+ iv
+w =
(z _ "' ) -a
(z - '" )iv
we have
+ u (z
_ '" ) i v
is bounded in
U
and the bound is independent
The question now is whether we can multiply each term in
(2.3) by a (different) bounded holomorphic function and still keep part b of the conclusion. p>l
For
p< 1
that is innnediate.
For
it is still true as can be seen by going through the proof
of Theorem 2.2.
(The crucial fact here is that the kernel in
Lemma 2.4 is positive and hence cancellation is not being used when combining the estimates for the individual terms. ) Corollary 3.6: Bl is an algebra. Tha tis, i f F,G are in then FG is in Bl and IIFGIl 1 < cllFlI 1 IIGII 1 B B B
Bl
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
247
Note that by the definition of F = I2f for some f in AlO •
Proof:
ros
we can write
in
F
Similarly we write
2 G = I g.
Thus we must show that
n2 (FG)
=
fG + 2 If Ig + Fg
is in
AlO .
A20 •
Hence the product is in
By the previous theorem,
note that, by Theorem 2. 10,
AlO
If
and
Ig
To see that
are both in fG
is in
AlO
G can be written as a sum of uniformly
bounded functions with summable coefficients.
Thus
Multiplication by a bounded function clearly takes The term Fg is in AlO for the same reason.
G f
is b ounde d. into AlO •
The individual terms in (2.11) are rational functions, all of the same degree.
Hence Theorem 2. 10 can be used to study questions
of rational approximation. on
For a holomorphic function
U we can measure how well
f
f
defined
can be approximated by rational
functions using the approximation numbers Rn (f) = inf{l!f - riiBloch ; r n
is a rational function of degree with poles in the lower half-plane}
or r
is a rational function of degree
n
with poles in the lower half-plane} •
In fact, either of these two sets of numbers characterize the spaces
BP Theorem 3.7: (a) space
f
is holomorphic in
.t P
U
,
O
f
is in
BP
i f and only i f
{R } n
is in the sequence
f
is in
BP
if and only if
{r n }
is in the sequence
.t P (b)
space
Suppose
248
R. ROCHBERG
Proof discussion:
f
If
BP
is in
then, by Theorem 2.10,
can be written in the form (2. 11) for some integer
b
be the non-increasing rearrangement of the numbers
{Ai}
number
~k
r
of
Since the Bloch norm of
to
f
the
k
maximum of the remaining ~k ~
the estimate sequence
Let
* {Ai} The
can be estimated by selecting as the rational appro-
ximation Ai.
f
[A k }
terms in (2.11) with the largest values
Ak
cA k* ·
can be estimated by the
(using Theorem 2.10 again) we get
Since the
tP
is in
f - r ~k
the sequence
are decreasing and the
~
is in
tP
The proof of the analogous estimate for the numbers
rk
goes
the same way as soon as we know which terms in (2.11) to use in the rational approximation. have small
BMO
The goal is that the remainder should
norm (as estimated using Theorem 2. 10).
This
requires the following Lemma 3.8: in k
U and
Suppose [A.}
0< p < ...
Suppose
{zi}
is a fixed
tP
is a sequence of numbers in
~
d-1attice
There is a
so that it is possible to renumber these sequences so that the
sequence with n-th term
II is in Proof discussion:
Let
indexed by
IJ.
be the measure
2 I: A. y. 6
Regard zi the points of the lattice as indexing a set of Car1eson boxes in
U •
Let
times the
be
IJ.
~
measure of the Car1eson box
• That is
It is fairly direct to show that the sequence the sequence
~
[Ai}
is.
[oil
is in
t P
if
Renumber the lattice so that the function
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
of
i
given by
249
max {ok : d (zi' zk) < I}
is non-increasing.
A
stopping time argument can then be used to show that the sequence with terms given by (3.9) is in t P • Details are given in [RS]. The proofs in the other direction, that functions with good rational approximation are in BP , requires the theory of Hankel operators.
It is completed by an appeal to Theorem 3.14.
(It would be interesting to free the second part of the proof from the theory of Hankel operators.
A related problem is to find
an appropriate analog of this theorem for harmonic functions in
Rn+ 1
+
.
)
Theorems 2.2 and 2.10 can be used to study the intermediate spaces (in the sense of real or complex interpolation theory) between various Besov and Bergman spaces. results in any detail.
We won't discuss those
The general idea is quite similar to that
in the previous theorem and it's proof.
The heart of the study
of intermediate spaces is splitting a function into two pieces with good control of the norm of the pieces in different normed spaces.
The previous proof was exactly of that type except that
we measured the "norm" of a rational function by it's degree. Standard interpolation theoretic techniques for the sequence spaces t P can be combined with Theorem 2.2 and Theorem 2.10 to yield most standard results about interpolation of the Bergman and the diagonal Besov spaces.
For instance, it follows easily from
Theorem 2.10 that the complex interpolation spaces between the Bloch space and BP are other BP spaces. USing the previous lemma one can also obtain the (slightly more subtle) result that the real intermediate spaces between BMO and BP are BP spaces. B.
Hankel Operators
In this section we use the decomposition theorems for the spaces BP to describe the Schatten ideal theory for Hankel
250
R. ROCHBERG
operators. [Po].
We will use SOme of the introductory material from
A different approach to these topics has been developed by
Peller,
[PI, 2, 3], [PH].
These references together with the survey
by Peetre [Pe2] give a good introduction to the theory of Hankel operators and Schatten ideals. Suppose
b
is a holomorphic function on
Hankel operator with symbol H2
H2
P
the
L2 (R,dx) ) given by
= Q( b f)
Here we write to
~,
is the map from the Hardy space
to its orthogonal complement (in ~ (f)
L2
b,
U.
(3. 10)
for the Cauchy (
Q= I - P
and write
= orthogonal)
projection of
for the complementary projection
onto conjugate analytic functions.
We will use the same name and 2 notation for the operator defined on L «O,oo),dt) by 00
~ (f) (t)
-
J b (s + t)f (s )ds
(Here, and throughout this section, form.)
(3.11)
o
denotes the Fourier trans-
These two operators are unitarily equivalent.
The equiva-
lence is implemented by the Fourier transform followed by the change of variables sending
t
to
-t.
(The equivalence of various points of view about Hankel operators uses the fact that the orthogonal complement of L2 H2.
H2
in
is the same as the space of complex conjugates of functions in For instance,
(3.10) to (3.11). variables,
that equivalence is used in the passage from In other contexts, such as functions of several
there is no natural substitute for this equivalence.
This causes problems in trying to decide what would be the "natural" extensions of the theory of Hankel operators to other contexts. ) The basic boundedness result for Hankel operators is that
~
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
251
will be a bounded operator if and only if more the
BMO
norm of
b
b
is in
BMO
Further-
is equivalent to the operator norm of
l1,. The general theme we will be developing is that small exactly if ~
is that
b
is smooth.
n
will be
A classical result of that sort
will be an operator of rank
rational function with
~
n
exactly if
b
is a
poles, all in the lower half plane.
Another result of the same general sort is that compact operator exactly if
b
l1,
will be a
has vanishing mean oscillation.
(Fqnctions of vanishing mean oscillation form a subset of consisting of relatively smooth functions. for instance, as those functions in
BMO
BMO
They can be described, for which the decomposi-
tion series (2.11) converges IN NOR~ ) The Schatten ideals (also called trace ideals and Schattenvon Neumann ideals) are classes of operators on Hilbert spaces with small ranges.
They are intermediate between the finite rank oper-
ators and the compact operators. operator
A
More precisely, for any bounded
(on a Hilbert space) and any non-negative integer
n
let s
n
=s (A)=inf[IIA-BIl ; n
B
is an operator of rank at most n}.
These numbers are called the singular values (singular numbers) of
A
If
A
is a compact positive operator, they are exactly
the eigenvalues of A
A
in decreasing order.
I
they are the eigenvalues of
the Schatten ideal
S
p
For a general compact
IA = (A*A) 1/2.
For
0< p <
00
is defined to be the set of operators for The definitions extend in the natural
way to maps between two spaces. The space S2
Sl
is sometimes called the Trace class; the space
is sometimes called the Hilbert-Schmidt class.
R.ROCHBERG
252 P~
For
1
these are Banach spaces.
p< 1
they are complete
quasi-normed spaces with (3.12) For any
p
p~1
For
the space
the
8
p
8
is a normed ideal;
p
norm of
A dominates the
diagonal entries of any matrix realization of [f} n
-t P norm of the A.
That is, if
is an orthonormal set then (3.13)
(In particular, if sentation of
p = 1,
the formal trace of any matrix repre-
A will be absolutely convergent ..• thus the name
"trace class".) Theorem 3. 14: 1'f and only if
For b
O
BP •
the Hankel operator
Hb
is in
8
p
Fur th ermore there i s an equ i va 1ence
The comments before the theorem show that for any function b of
in Hb
BMO,
there is an inequality between the singular values
and the approximation numbers of
b.
sn(Hb)-;;crn(b) •
Hence the remaining implication in part b of Theorem 3.7 follows from Theorem 3.14.
.
The completion of the proof of part
a
of
Theorem 3.7 is a bit more intricate, see [8] • Proof Discussion:
There are two ideas which relate the decomposi-
tion theorems to Hankel operators.
First of all, the individual
terms in the series (2.11) generate Hankel operators of especially simple form. of
H2
8econd, the building blocks used in the decomposition
in Theorem 2.8 come close to diagona1izing Hankel operators.
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
253
We will first look at how these two ideas give the two halves of the theorem for Suppose b (z) = !: A.
=1
is in
b (
1
p
Bl
By Theorem 2. 10 we can write
b
as
The Fourier transforms of the individual
_) z - z. 1
terms are quite simple: -1 A -iz.t (y. (z - z. ) ) (t) = y . e 1
(t> 0) .
1 1 1
Thus the Hankel operator corresponding to an individual term is the operator which takes
to
f
H. f : 1
S y.e izi(t+s) f (s ) ds a>
o
1
This is the one dimensional operator which takes -iz.t 1
with
f
Now note that the norm of
to
(f,ei>;i
and hence
also the operator norm of this one dimensional operator, is independent of
Thus the sum (2.11) generates a sum of one
z.. 1
d~ensional operators with individual norms at most
the operator norm and the operators we have
IIHblll =
S
p
cIA.I. 1
Since
norm are the same for one-dimensional
II !:AiHi ll
-;;!:
lA)
IIHilll -;; c!: IAil -;; cllbll 1 B
which is the desired estimate. When if
b
p< 1
is in
BP
the lack of convexity works with us.
That is,
then we can argue the same way and then use (3.12)
in place of the triangle inequality to conclude that
~
is in
S
p
Suppose now that d-lattice
[zi}
~
with small
is given. d
and let
the corresponding building blocks for
H2
fj =
Pick and fix a 3/2 - - 2 be yj (z - Zj)
in the sense of Theorem
R.ROCHBERG
254
2. B.
The functions
fj
are not an orthonormal set but they are (That follows
the image of an orthonormal set under a bounded map. easily from Theorem 2.B.)
Hence (since the
estimate (3.13) also holds for the work directly with the picture.
f.
fj •
S
p
are ideals) the
However, rather than
we will work with the Fourier transform
J
This has the advantage that the domain and range space
of the operator are the same.
It also gives us an opportunity to
correct some arithmetical errors which occur in [Rl].
We wish to
To compute this we first compute that
(t> 0) •
Thus A
~
A
(R. f., f.> =c -0 J
J
3
II b(s+t)y.J tse
-iz.t -iz.s J e J dsdt
When we change the variables
t
w= s + t
integral this gives
and evaluate the 00_
I
C
t
and
s
to new variables
t
and
•
3 -1Z'W b (w)w e J dw. A
o
This is the Fourier transform formula for computing the third derivative, hence
A...
3 J
(Ii f. , f. > = c y. b -lb J
J
(3)
(3.15)
(z . ) J
Thus the sum we are considering is a Riemann sum for the integral we wish to estimate.
f. , f.) I. -0 J J
c t I (R. all
Specifically
t ly~b3(z.)! J
J
Taking the supremum of the left hand side over
d-lattices gives the required estimate:
~
IIbllB 1
cI: I(~fj' fj>! ~ cll~lIl' CRl] for
&
= 1/4.
be replaced by
(This is the argument on pg. 917 of
Severa 1 places in tha t argument,
2&.)
&
should
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
255
This argument extends unchanged to all
p> 1.
The only new
ingredient we need is to note that (3.13) is valid for those Suppose now tha t
p> 1
Hb
want to show that
and the symbol
is in
Sp
b
is in
p.
BP •
We
This follows from interpolation
theory applied to the map from symbols to Hankel operators. know that that map takes functions in
Bl
We
to operators in
Sl
(from the first part of the proof) and we know that the map takes functions in
BMO
to bounded operators.
Interpolation theory
then insures that the map takes the (appropriately defined) spaces between
Bl
and
BMO
to the corresponding spaces between
and bounded operators.
Sl
It is a result from the theory of Schatten
ideals that these latter spaces can be taken to be the spaces
S
P
We noted in the previous section that Lemma 3.8 can be used to show that the spaces intermediate between are the spaces in
BP •
BP
1< P<
00.
~
Hence
is in
Bl
and
Sp
if
BMO b
is
(T:his is an appeal to "real interpolation theory".
could instead use "complex interpolation theory".
One
That approach
uses simpler results from interpolation theory but uses a more complicated map from symbols to operators.
A proof without inter-
polation theory can be given using Theorem 3.7). Suppose now tha t is in
S
p< 1
and we know tha t the opera tor
We wish to show that
p
valid for
p< 1
b
is in
BP .
Hb
(3. 13) is not
and so we must work directly with the definition.
This proof is more complicated and we only outline the ideas. Start with a lattice and functions
k - 1/2 --k fj=Yj (z-Zj)
by Theorem 2.8, from a sort of basis for
H2.
we need
IIp.
fj
k
to be large, of the order of
were orthonormal.
which,
For this argument Suppose that the
(They are not but, by Theorem 2.8, they are
boundedly equivalent to an orthonormal set and hence are an acceptable substitute in what follows.) ~
relative to the
fj
by
(a ij ) .
Denote the matrix elements of A calculation similar to
the one which led to (3.15) but with general indices shows that
256
R.ROCHBERG
an analog of (3.15) holds (with 3 replaced by
2k - 1).
Thus
the diagonal elements of the matrix are exactly the quantities we wish to estimate---high order derivatives of of
y.
b
times powers
We would like to show that
I I: y71 k - l b (2k -
1) (z.
1
)I P
=
I: la .. IP < cliH. liP 11 = -1> P
(3.16)
If this matrix were diagonal then (3.16) would be a consequence of the definition of
What is true is that the matrix is
S
p
nearly diagonal in a useful sense. where
D is the diagonal part.
Write the matrix as
D+R
If we could show that the
norm of the off diagonal remainder
R
S
p
satisfied (3.17)
for a very small constant
IIbli p B
IIDli s
c
,
then we would have
p
which gives the required control of
b.
(3.17) is true if the lattice constant, The reason is that estimates for the
S
p
d,
is very large.
norm of the matrix can
be obtained by combining the norm estimates for the one dimensional operators corresponding to the individual matrix entries. this proceedure is carried out for particular choices of
When b,
the
resulting estimate is a Riemann sum for an integral similar to the one estimate. in Lemma 2.5.
If attention is restricted to the off
diagonal elements of the matrix then the corresponding integral is of the same general sort but the domain of integration is restricted to the set of
z
integral controls
with c
,
d (z,,) > d.
Since the value of that we may make c' sma 11 by making d
large. We cannot choose the lattice constant
d
to be large in
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
Theorem 2.B.
257
However, we can take a lattice with small
express it as a union of lattices with very large this has the effect of splitting many subs paces.
2
H
d.
d
and
Roughly,
as a direct sum of finitely
The Hankel operator which is being studied can,
in effect, be realized as a direct sum of operators of the same general sort acting on each of these subspaces.
The considerations
of the previous paragraph apply to each of those operators.
We
can then combine the estimates without loss because the operator norm estimate for a direct sum of operators is the maximum of the individual estimates. The details of this proof are in [R1], [S], and, in an improved form, in [RS].
Alternative proofs are given by Peller [P1,2,3]
whose proof of Theorem 3.14 for
p= 1
(by quite different tech-
niques) was the inspiration for much of the recent work in this area. Analogs of Theorem 3.14 hold for some operators closely related to Hankel operators. we define the operator defined on
(0,=)
~~f(X)
=
For instance, for any complex
~,~
~~ to be the operator acting on functions
given by
J= s~t~(s+t)-~-~ o
b(s+t)f(t)dt
D~ + ~ c = b. These opera tors can c also be viewed as Hankel type operators acting between potential This opera tor is
D~H D~ wi th
spaces (i.e. spaces of the form
D~2 ).
Alternatively, these operators can be regarded as Hankel type operators on the Bergman spaces A2Y • That is, one can define operators on the A2Y using formulas similar to (3.10) but starting with
f
in the Bergman space and using the Bergman projection.
(Such operators were studied in [CRW].) Fractional integration 2Y 2 gives a unitary equivalence of A and H and hence can be used to pull these operators over to .~.
When this is done (by
258
R.ROCHBERG
straightforward Fourier transform calculation) the resulting operators are of the form
~~.
The techniques we have been discussing (and also those of Peller) extend fairly directly to these more general operators if some restrictions are put on ~
and
a
~
and
(What happens for
a
outside that range is a bit of a mystery. )
Theorem 3.18:
Suppose
min(l/Z,l/p»O.
~~
0< p < <Xl is in
and that
min (Re a, Re
i f and only if
Sp
b
~)
+
is in
B
P
Furthermore the norms are equivalent. A class of operators which has been studied a lot recently are the operators defined on plication by a
LZ (lRn)
as commutators between multi-
(relatively) smooth function and operators given by
(possibly singular) intergral kernels.
The basic philosophy is
that the smoothness of the function used in multiplication helps balance the singularities of the kernel and the resulting operator will be tamer than the
integral operator alone. LZ (lR) ) the projection
In one dimension (i. e. on onto
HZ
P
LZ
of
can be taken as the fundamental singular integral operator.
This leads to a close relation between Hankel operators (and, more generally, the
~~)and
commutators.
We will say that a function P(b)
b
BP
is in the space real
and the complex conjugate of
(I-P)(b) BP •
holomorphic type) are in the space
if
(which are both of
We will use the same
notation for a function and the operator of multiplication by that function. where
D
We are interested in the commutators is differentiation and
the so called Calderon commutator. bounded on LZ if b is in BMO
DB = b.
[b,P)
and
[B,P)D
The second opera tor is
Both of these operators are (see
[CMl, Z)).
e. g.
We discusS
the boundedness in Section D. Theorem 3.19: if
b
is
Suppose
in the real
O
[b,P)
[B,P)D
is in
is in
S
p
S
if and only p i f and only if
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
b
BP
is in real
~:
259
The basic idea is that these commutators are direct sums
of operators of the type we have been discussing. [b, P] f
Set
[b,P]Pf+ [b,P]Qf
-[b,Q]Pf + [b,P]Qf
-Q (bPi) + P(bQf)
-Q «Qb ) Pf) + P «Pb )Qf)
Q= I - P •
Thus the commutator is a direct sum of two Hankel operators. Qb
is the
b
of (3.10».
the two pieces are. direct sum of two
Hence [b,P]
is in
S
(Here
if and only if
p
The Calderon commutator is essentially a KO,l
operators.
(Analogous results for commutator operators acting on L2(Rn) are given in [JW], and [JP].
The techniques used there are quite
different. )
c.
Interpolation of Values in Bergman Spaces In this section we describe sets
S
in
U on which the values
of functions in the Bergman spaces can be specified in quite freely. More precisely we want to describe those for any function defined on
S
S with the property that
and satisfying natural size con-
ditions, there is a function in the appropriate with the given function on
A~
which agrees
S.
For background, and for contrast with the results for Bergman spaces, we begin by recalling Carleson's interpolation theorem for bounded analytic functions. CD H
Denote by
CD H
the space of bounded analytic functions on
U.
is normed by the supremum norm. A sequence
[zi}
is called an interpolating sequence
HCD ) if for any bounded sequence of values CD to find a function f in H so that i = 1,2,3 •••
(for
it is possible
(3.20)
R.ROCHBERG
260
It is a consequence of the open mapping theorem (applied to the map which restricts functions in
H=
to bounded functions on
S)
that if (3.20) can always be solved then it can be solved with norm estimates.
That is, there will be a
c
so that
f
can be
choosen with (3. 21)
The boundedness of of f
f
in
f
implies estimates for the derivative
and hence estimates on the modulus of continuity of U • Hence, if there is a function f .. in H= of U
inside
1.J
norm at most
c
which takes the value one at some Z.
and is zero
1.
at some
Zj'
then
zi
must be separated from
Zj.
One can
make this informal idea quantitative using Cauchy's theorem or using the Schwarz
lenma and the invariant geometry.
In this case
we do the latter. The Schwarz-Pick lenma (which is the conformally invariant form of the Schwarz lemma) says that an analytic map of a disk of radius
c
is a contraction when both
are given their hyperbolic metrics.
U and the disk
When we apply this to the
function
f ij
we obtain a lower bound on
only
c.
Thus a necessary condition on the set
on S
is
s-separa ted for some
s
>0
H=
interpolation. ident~cally.
which depends S
is that (3.22)
S
that are much too
For instance, if a bounded analytic
function is zero at the points of a it must vanish
d{zi' Zj)
•
This condition by itself allows sets thick for
U into
d-lattice (for any
d) then
The additional condition that is
needed is tha t is a Carleson measure. Carleson's theorem is that (3.22) and (3.23) together are the necessary and sufficient condition on (3.20) can always be satisfied.
S
to insure that
This, and much more on
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
261
interpolation problems for the Hardy spaces,
[G].
are
We now consider the same sort of problem for the Bergman spaces. If
f
For convenience we start with the case 10
is in
A
and
z
is in
p = 1,
U then the size of
ex. = 0 •
f
can be estimated using the area mean value theorem for a disk in of
y
2
U centered at
z.
2
z
~uclidean)
The area of the disk is of the order
and we obtain the estimate
exponent
at
If (z) I ~ cllflly -2.
(The
is sharp as can be seen by considering the individual
terms in the decomposition formula (2.3).) consider the normalized quantities
This suggests that we
2
Yif(zi)
in the interpolation
problem. For
z
in
f , (z) = c
U (
Iz-c
f(C)(C -
z) -2
dC.
=y/2
When this is combined with the estimate for
f
we find
If'(z)1
~ cy -3I\fll.
Using the estimates for f and f' we conclude that the quantity y 2 If(z)1 is a Lipschitz function on U; that is (3.24) in
U.
Suppose now that we have a sequence U
S = [zi}
of points in
We describe the interpolation problem in terms of the operator
T which takes
f to the sequence Tf = [y2 f (z )}. We have seen n n that T maps AlO to bounded sequences. If we want the image of a bounded set in AlO to contain all of the characteristic functions of singletons in
S
then, but (3.24), (3.22) must be
satisfied. If (3.22) is, in fact, satisfied, then the estimate of the size of
Tf
for
f
10
in A
ball of Euclidean radius
can be improved. Let /lOO Yi centered at zi.
Bi
be the
262
R.ROCHBERG
'$ C
SIu
1fl
lIfll •
=
Here we are using the fact that the set clude that the the area of
B
n
B
is scattered to con-
are disjoint and we are using the fact that 2
is comparable to
n
S
Yn •
This gives us half of the following Theorem 3.25: If S satisfies (3.22) then T is a bounded map from AlO into the sequence space t l If s in (3.22) is l large enough, then the map is onto t In fact if s is sufficiently large then the map a bounded linear map from
tl
Proof:
T
We have seen that
T will have a right inverse
to
such tha t
A 10
is bounded.
R ,
TR = I .
Suppose now that
S
is given and that (3.22) is satisfied for some =
3
--3 J
(2i) y.(z-z.) • J
Hence the map
R
o
It is almost true that . 1n
of
'\.1 .-
Ro ({a.}) = L a.f. 1 1 1
to
good approximation to the desired
R
that the operator norm of
is small.
lI(I-TR o )({a.})lI = L 1 = L
< c
1- TR
la k Ia k
o
More precisely we show
- L a.y. (2i) j J J
I I
L La. k j Fk J
j
Pick
{a i }
(TR 0 ({a,})(zk)l 1
c L laJ.IY J. (
is a
YJ.Y k2
!: k
#j
3
2 -31 yk(zk - z.) J
I zk - -zJ' 1- 3 I zk -
Zj
I
3
in
t1 •
263
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
Now, in each integral make the change of variables of (z +xj)/Yj.
< Because
z
to
We can now continue the estimates with
c
J I, + il-~ dV
is finite we can insure
J
c
is
d(C,i»s less than one by selecting s sufficiently large.
It follows that
TR = (1- (r,.-TR» o 0
is the required
is invertible.
R=R (TR )-1 0
0
operator and the proof is finished. Note that the use made here of the hypothesis that large is similar to the use made of the similar
s
• hypothesis
be in
deriving (3.17). Only a little has to be added to extend this result to all Set
Bergman spaces. sequence in in
U
tP
•
r=(2+o,)/p.
Let
T
r
Suppose
Suppose
S=
[z.} J
is a given
be the map from functions holomorphic
to sequences given by
Theorem 3.26: then
U.
T (f) = [y~ f (z .)} r J J
O
-1<0,
If
S
satisfies (3.22)
T is a continuous map from APo, to the sequence space r If the constant s in (3.22) is large enough (i.e.
s> s (p,o,) ) then the map is onto. right inverse
R
Proof discussion:
for
In this case there is a bounded
T
We discuss briefly how the steps in the previous
proof can be extended. To show that
T
r
is bounded for
p>l
we use the area mean
value theorem (as before) together with Holder's inequality (to get to an estimate involving integrals of
I f\ P
.)
When
We follow the same pattern but must also use the fact that
p< 1
R.ROCHBERG
264
If(z)I P
is subharmonic to control
If(zk)I P
by the appropriate
integral. To define the map
R
o
we select
with a choosen large enough so that Theorem 2.2 applies.
R
is
o
constructed as before and the previous proof works almost without change for TR
o
p< I.
if
is even bounded.
p> I
then it is no longer clear that
The problem is exactly the discrete analog
of the problem which arose in proving Theorern2.2---we had to show that the right hand side of (2.3) was in the required space. problem is solved using the same tools.
The
The operator being studied
is a discrete analog of the operator in Lemma 2.4. have a Riemann sum for the integral operator.)
(That is, we
The boundedness
of the operator is proved by appealing to a discrete analog of Lemma 2.6.
The estimates needed in that lemma corne from a discrete
analog of Lemma 2.5. To show that the operator
I - TR
o
is small, we examine the
proof that it is bounded and find that one of the controlling factors is an integral which, after a change of variable, reduces to an integral over the se t of
z
for which
before, it is here that the choice of large
d (z, i) > s s
As
is made.
This proof is carried out in detail in [R2] for general Bergman spaces and a refined version is given in [RS]. ideas are used by Amar in [A].
Similar
Luecking obtains these results and
others using a different approach [LI,2,3].
He then uses his
results to obtain decomposition theorems. This theorem shows that the values of a function in an at different points of a thin lattice (i.e. a large
d) are rather independent.
is thick (i.e. if totally.
d
AP~
d-Iattice for a
In contrast, if the lattice
is small) then this independence fails
MOre precisely, if the lattice is thick then the values
of the function on any finite subset of the lattice are determined by the values on the rest of the lattice.
To see this it suffices
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
to show that a function in vanishes identically. Theorem 2.2.
If
f
265
APQ which vanishes on a thick lattice
That is a consequence of the proof of vanishes at all the lattice points then the
first approximation to the decomposition (2.3) will be identically zero.
However, the proof shows that the first approximation is a
good approximation in norm.
Hence
f
must be the zero function.
In particular, the condition in the previous theorem that
s
be
large is certainly necessary. We noted before that formally some results for Hardy spaces are the limiting case of the results for Bergman spaces as goes to
-1.
That is not true here.
spaces is independent of
The result for Bergman
and the analogous
Q
Q
re~ult
for the
Hardy spaces is that conditions (3.22) and (3.23) are necessary and sufficient for inverse.
T
to be a bounded map with a continuous right
When the proofs are looked at in detail it is possible
to see exactly where the analogy fails.
proof of the theorem we needed to know that if
[A k }
is square summable.
p =2 •
Suppose ~
Akfk
is in
That conclusion follows i f the
matrix of inner products «f j , f k » is a bounded operator on the sequence space ~2 By direct computation that matrix is
.
for positive if
[zi}
&.
The proof shows that such a matrix is bounded
is a separated sequence.
However, if
&
=
0
then an
additional condition is needed for the matrix to be bounded, [z.} ~
must also satisfy (3.23). detail in [Ko].) fails at (A k } .
b= 1
(This relationship is developed in
This is analogous to the fact that Theorem 2.8 unless a further condition is put on the numbers
(Such conditions will be used in the next section, see
Lemma 3. 10. )
R.ROCHBERG
266
It would be very nice to have an analog of the previous BP , or
theorem for the spaces Bloch, context,
BMO
spaces. BMO
BMO
(However, in this
should be considered as part of the scale of Hardy
Results about interpolation of values by functions in
are given by Sundberg in [Su]).
The methods we have described
are better suited for studying the derivatives of the functions in those spaces than the functions themselves.
For instance,
using a variation on the previous proof it is possible to describe the values that the derivative of a function in BMO or the derivative of a function in a Hardy space HP can take on as-separated set if
s
is large.
This gives results for functions in the tent
spaces of [CMS2]. D.
Bounded Operators, Local Estimates, Higher Order Estimates We now sketch some extensions of the ideas of the earlier
sections.
The details are in [RS].
We will rely more on back-
ground information from [G] about Carleson measures and related things.
We will use the real variable analogs of the earlier
results.
Those involve lattices in the upper and lower half-planes.
First we discuss the boundedness of the commutator operators of Theorem 3.19.
For more on the background and role of such
operators see the papers of Coifman and Meyer [CMl,2], where, among other things, there is an earlier proof of the following theorem. Theorem 3.27: L2 (R).
If
and on
BMO
the
BMO
If B'=b
is in is in
BMO BMO
then then
[b,P] [B,P]D
is bounded on is bounded on
L2(R)
In each case the operator norm can be estimated by
norm of
The first
b
L2
b. result follows from the second since the first
operator is essentially the difference between the second one and its adjoint. The projection operator is a linear combination of the
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
267
identity and the Hilbert transform. 0>
=c
([B, P]D)f(x)
J
b (u) - B (x, u)
x-u
-0>
with
Integration by parts gives
B(x,u) = (B(x) - B(u»/(x - u)
parameter
b=2
f (u)du
(3.28)
If we use Theorem 2.10 with
to decompose the function
b,
we find (using
partial fractions) that the kernel of (3.28) can be written
b (u) - B (x, u)
c I:
x-u
A. (x - z. ) ~
(u - z.)
~
2
~
Thus [B,P]Df = c I: A.(f,g.) f. ~
~
~
and
with the ftinctions be in
L2
By Theorem 2.8
are weakly orthonormal, that is
if the
ai
are square summable.
same property we would only need that
Ai
If
I: a.g. ~
{f i }
will
~
had the
are bounded to conclude
that the operator given by (3.29) is bounded.
However
not that nice (and that conclusion would be wrong).
{f.}
is
~
What is true
is Lemma 3 30: sequence
Suppose
{Ai}
{zi}
is a
is such that
d-lattice for some
~ y.A:6 ~
[Aif i }
~:
and compute
I:
\LI 2 ~
g
Since
y.g(z.) ~ ~
I: 1Ai\2 Yi6z.
L2
g
in
is a Carleson measure
~
this last quantity is dominated by for all
If the
zi is weakly orthonormal.
~
then the set of functions Take a
d.
is a Carleson measure
H2
cllgll 2 •
Having this estimate
is equivalent to the required estimate.
case of the theorem follows from this. We now consider the
BMO
boundedness.
It is enough to
The
R.ROCHBERG
268
consider the case when
B
is conjugate analytic.
In that case
h = [B, P ]Df
Thus
h
is conjugate analytic and is determined by
show
h
is in
Pick
g
with
BMO g
P(f).
by showing it pairs with functions in HI.
in
< c!: II.1 . I Yi If' (z. ) I I g (z. ) I 1 1
I (h, g) I
The first factor is controlled using the Carleson measure condition on the
Ai
and the fact that
controlled by the of
!:
I
H
norm of
y~If'(z.)126 1 1 z.
g
is in g
HI.
The second factor is
and the Carleson measure norm
That measure is a discrete version of
1
If' (z) 12 ydxdy
which is a Carleson measure because
f
is in
BMO.
The operators just considered are the linear terms in the multilinear expansion (i.e. Taylor series) of the weighted norm inequalities for the Cauchy projection and of the Cauchy integral on chord-arc curves.
With more effort, the same techniques---
decomposition of the symbol of the operator, partial fraction analysis of the kernel, and weak orthonormality results obtained using Carleson measure estimates---can be used on the quadratic terms in both
e~pansions.
That is, the same ingredients can be
used to establish the boundedness on L2 with b in BMO and of [A, [A,PD 2 ]] if
of the operator
[b, [b,P]]
A'
Other
is bounded.
trace ideal results for operators related to the Cauchy integral are in [S2]. In the proof of the decomposition theorems there is a close
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
relation between the value of (2.3) or (2.11).
f
269
at
zi
and the numbers
Ai
in
Similarly in the proof of the theorem on inter-
polation of values there is a close relation between the value which the function
f
is to take at
constructed function near
zi'
z.
1
In the previous sections we didn't
try to make this local control precise. norm control.
and the size of the Instead we settled for
It is possible however to extract the local estimates
from the proof.
The key is to estimate the spreading effect when
the crucial approximation step is iterated.
One way to do this
is to get good estimates on the high powers of the operator in Lemma 2.4.
T
Here is an example of the type of control which is
possible. Lemma 3.31:
Suppose
is given.
&>0
possible to choose the numbers BMO
function
f
so that
Ai
In Theorem 2.10 it is
in the decomposi tion of the
IA.I
1
1
1
Here
G
is the
integral operator (Gg)(w) =
JJu
Ig(w)1
':1 3 _ &
I
w- z
Approximation numbers (the
s
n
(~+ y
1& ) dxdy • v
) of Hankel operators and
related operators can be estimated directly in terms of the which appear in the decomposition of the symbol.
Ai
Hence this lemma
and its variations allow very precise control over approximation numbers.
We will give an example involving products of Hankel
operators in a moment. Similarly, there is the following refined version of Theorem 3.26.
Theorem 3.32: set.
Suppose
&> O.
Suppose
Suppose a sequence of values
Solve the interpolation problem
,where of center
b.
1
f (zi)
{zi}
is an
s-scattered
is given and we wish to
=b i '
Let
b (z)
=I: b iXi (z)
is the characteristic function of the hyperbolic disk and radius
1 •
If
s
is sufficiently large then
270
R.ROCHBERG
a sufficient condition that the interpolation problem have a solution is that for some (and hence any) finite.
In that case the function
f
z
in
U,
Gb(z) be
can be choosen to satisfy
If (z) I < c Gb (z) • Theorem 3.26 is a special case of this because operator
G
the
(or other operators for which similar results hold)
is bounded on the various
AP~.
The point of both proofs is that for small G dominates the operator tion when iterates of
~ C~n.
Thus
T are combined.
C the operator
G controls the situaThe fact that
G has
the required properties is obtained by direct computational estimates. We now describe a singular value estimate for products of Hankel operators.
We will only indicate some main themes of the
proof. We consider Hankel operators as maps from of functions in
H2.
~
That is,
that we have two such operators, and
b2
H2
to conjugates
is given by (3.10).
HI
and
H2
Suppose
with symbols
bl
2 * (which maps H HlH2 In addition to its intrinsic interest, this operator
and we wish to study the composite
to itself).
arises in the study of Toeplitz operators as the semi-commutator HI*H2 = T"b b - T"b Tb 1 2
1
where the
TI s are Toeplitz operators.
2
With Hankel operators and with commutators a basic theme was that smoothness in the symbol led to smallness of the operator (i.e. small singular values).
Here the basic theme is that the
smallness of the product operator is controled by the product of measures of smoothness for the two factors. [zi}
U• b (x)
by
is a
d-lattice in
The derivatives of
U and b
at
(the boundary values of Yi.
b
That is, suppose
is a function holomorphic in measure the smoothness of near
and on a scale given
This informal reasoning suggests that we try to control
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
271
the product operator with the numbers
b' (zl)b' (z2) •
Although this general philosophy is correct, in order to get the proof to work we use a different measure of smoothness, one based on mean oscillation.
Also, the estimates on the singular
numbers involve the products of smoothness estimates summed over Carleson boxes. Suppose p>2 U
and
and
z
d
is a positive integer and
n>pd+ I and
w
nomial of degree
For a function in
d
U of
we let b
b
nand
satisfy
which is holomorphic in
P(b)(z)(w)
at the point
(P(b)(z)(w)=b(z)+b'(z)(w-z)+ •.• ).
p
z
be the Taylor polyevaluated at
For a point
w
z=x+iy
in
U we measure the smoothness of the boundary value function at the point
x
and the
sc~le
y
by the number
osc(b,z)
b(t) given
by CZ>
osc (b, z)
(J -CZ>
n - I Ib (t) - P(b) (z) (t) P _Y'---_ It -
I
z,n
dt) IIp •
Roughly, this quantity measures the oscillation of Over the Carleson box centered at
z
b(w)
Various spaces of distri-
butions on the line which are characterized by mean oscillation conditions can also be characterized by the size of these numbers as
z
varies over an
0< p < co
r-Iattice.
In particular the spaces
can be characterized that way.
Suppose tha t
{z.} 1
is a
BP
(See (11.9) of [RT].)
r-Iattice in
U.
Define
D.1
Let
*
tD.}
~eorem
1
be the non-increasing rearrangment of
3.33:
If the lattice constant
there are constants
c
and
K
so that
r
tD.}. 1
is small enough then
R.ROCHBERG
272
n=1,2,3, ••.• To understand this theorem a bit better we first restrict attention to the case osc (b, zk)
2
•
b
If
b1 = b2 •
D n
is then a sum involving
is assumed to be in
ation numbers are bounded.
BMO
then these oscill-
However, the fact that the oscillation
to be fini teo n Hence the theorem does not give the sharp boundedness criterion. numbers are bounded is not enough to force
D
On the other hand, if b is in BP then the oscillation numbers will be in t P and that is enough to force the D to be in 2 t P/ This gives the conclusion that is in nsp This
Eb
argument works for all
p
and can be used to replace the inter-
po1ution argument in the proof of Theorem 3. 14. Suppose now that there are two different operators.
Just
as we could not recover the boundedness criterion for a single operator, we do not recover the very nice results ofAx1er, Chang, Sarason, [ACS] and Vo1berg [V] which give a necessary and sufficient condition for
H*H2 to be compact. There are, however, 1 conclusions to be drawn. If b i is in l i then osc (bi' zk) is in
t
P. and hence the product operator is in
1.
S
P
l/p =
with
We didn't really need the theorem to get this result (because an analog of Holder's inequality holds for the Schatten ideals) but we can also get a localized version.
That
is, suppose th8t each point of the (extended) line has a neighborhood in which the boundary values of those of
b2
in
opera tor is in
B
S
P2
are locally in
l/p = 1/P1 + 1/P2
with
P1
and
P2
and
then the product
(Of course one must define "locally" and
P
the various estimates must be un form. ) that the indices
b1
The innovation here is
may vary from point to point.
273
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
Another consequence of the theorem is that it gives boundedness and Schatten ideal criteria which apply even when one of the operators is unbounded. p
For instance, if
then the oscillation numbers of
a way that can be estimated.
If
bl b2
bl
is in
LP
for some
will be unbounded, but in is smooth enough then its
oscillation numbers will be small and can balance this unboundednesS.
The theorem allows this to be made precise and, as before,
to be localized. Here are some of the ideas of the proof of the theorem. ~ =
suppose tha t we only had one opera tor, that f
b
is conjugate analytic.
H
explicitly
1
to each term.
in
H2
and decompose
1
We can
By direct calculation
b(x) -P(b)(z.)(x)
m - 1/2 Yi
~(
f
We may suppose
m - 1/2 --m f = I: A. y . (z - z i ) .
according to Theorem 2.8 as
apply
Pick
H.
First
1
Now the point is that this vector is the number
osc (b, zk)
times a vector with good orthogonality properties (more precisely, the same type of orthogonality properties as the functions in the proof of Lemma 3.30. tion shows up.)
f.1
That's why a Carleson measure condi-
The exponent
p
in the definition of
osc(p,z)
is to allow for the use of Holder's inequality in the demonstration of this orthogonality.
p>2
is required so that the conjugate
exponent will be less than two.
That allows quadratic estimates
on other terms to be obtained using a maximal function. To use the estimates on the oscillation numbers we partition the lattice into two sets.
A finite set of controlled size gives
rise to the finite rank operator which does the approximation. norm estimate for the oscillation numbers associated with the complement gives an operator norm estimate for approximation.
H
minus the
To do the analysis of the product operator, the
A
R.ROCHBERG
274
second operator is applied to the coefficient functional which sends
to
f
A. (f) 1
(Note that this functional is in the domain
HI* . ) I t is in the analysis of that coefficient functional that estimates are needed of the spreading effect produced by
of
iteration of the approximation operator.
Those estimates are
similar to the ones needed for the preceding two theorems. 1.
Supported in part by the N. S.F.
REFERENCES [A]
Amar, E. "Suites d'interpolation pour les calsses de Bergman de la boule et du polydisque de en". 1978, Canad.J. Math 30, pp. 711-737.
[ACS]
Axler, S.; Chang, S.-Y.A; and Sarason, D. "Products of Toeplitz operators". 1978, Integral Equations and Operator Theory 1, pp. 285-309.
[BCFJ]
Beckner, W.; Calderon, A.P.; Fefferman, R.; and Jones, P.; eds. "Conference on harmonic analysis in honor of Antoni Zygmund". 1983, Wadsworth Inc., Belmont, Ca.
[B]
Bui, H.-Q. "Representation theorems and atomic decomposition of Besov spaces". 1984, manuscript.
[CGT]
Chao, J. -A.; Gilbert, J.; and Tomas, P. "Molecular decompositions in HP theory". 1980, Supp. Rend. Circolo mate, Palermo, ser II, pp. 115-119.
[CGT] in
"Molecular decompositions and Beurling spaces HP theory". 1982, preprint.
[C 1
Cohen, G. "Hardy spaces: Atomic decomposition, area function and some new spaces of distributions". June 1982, Ph.D. Dissertation, U. of Chicago, Chicago, II.
[CMl]
Coifman, R.R. and Meyer, Y. "Fourier analysiS of multilinear convolution Calderon's theorem, and analysis on Lipschitz curves". 1980, Lecture Notes in Mathematics 779, pp. 104-122.
[CM21
•
"Le theoreme de Calderon par les methods de 1979, C. R. Acad. S.... , Paris, sei.'les
-v-a-r-i-.:l-:-b-::'l-e-r~clle". A~
289.
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
275
[CMSl]
Coifman, R. R. ; Meyer, Y.; and Stein, E. M. "Un nouvel espace fonctionnel adapt~ a l'etude des operateurs definis par des integrales singulieres". 1983, Lecture Notes in Mathematics 992, pp. 1-15.
[CMS2]
"Some new function spaces and their applications to harmonic analysis". 1984, preprint.
[CR.]
Coifman, R.R. and Rochberg, R. "Representation theorems for holomorphic and harmonic functions in LP". 1980, Asterisque 77, pp. 11-66. Correction to [CR.], in preparation.
[CR.2]
[ew]
Coifman, R. R. and Weiss, G. and their use in analysis". 83, pp. 569-645.
"Extensions of Hardy spaces 1977, Bull. Amer. Math. Soc.
[F]
Frazier, M.
[G]
Garnett, J. "Bounded Analytic Functions". Press, New York.
[JP]
Jan~on, S. and Peetre, J. "Higher order commutators of singular integral operators". 1984, manuscript.
[JW]
Janson, S. and Wolff, T. "Schatten classes and corrunutators of singular integral operators". 1982, Ark. Mat. 20, pp. 301-310.
[J]
"On the decomposition of Jodeit, M. into humps, in [BCFJ].
[Ko]
Koosis, P. "Lectures on HP spaces". 1980, L. M. S. Lecture Notes Series 40, Cambridge Univ. Press, London.
[KJ
Kra, 1. "Automorphic forms and Kleinian groups". W.A. Benjamin Inc., Reading Mass.
[Ll J
Luecking, D. "Closed range restriction opera tors on weighted Bergman spaces". 1984, Pac. J. Math 110, pp. 145-160.
[12 J
"Forward and reverse Carleson inequalities for functions in Bergman spaces and their derivatives". Amer. J. Math., to appear.
[L3 J
"Representation and duality in weigh ted spaces of analytic functions". 1984, manuscript.
B. Jawerth, 1984, manuscript.
L 1 (R) 1
1981, Academic
functions
1972,
276
R.ROCHBERG
[M]
Merryfield, K. "The molecular decomposition of certain Bergman spaces". 1984, manuscript.
[Pe]
Peetre, J. "New thoughts of Besov spaces". 1976, Duke University Math Series 1, Dept. Math., Duke U., Durham, NC.
[Pe2 ]
"Hankel operators rational approximation and allied questions of analysis". 1983, Second Edmonton Conference on Approximation Theory, CMS Conference Proceedings 3, Amer. Math. Soc. Providence, RI, pp. 287-332.
[PI]
Peller, V. V. "Smooth Hankel operators and their applications". 1980, Soviet Math. Dok1. 21.
[P2 ]
"Hankel operators of the class ~ and their applications (rational approximation, Gaussian processes, majorization problems for operators)". 1982, Math. USSR Sbornik 41, pp. 443-479.
[P3 ]
"Vectorial Hankel operators, connnutators, and related operators of Schatten-von Neumann class ~". 1983, Integral Equations and Operator Theory 5, pp. 244-272.
[PH]
Peller, V. V. and Hruscev, S. V. "Hankel operators, best approximations, and stationary Gaussian processes, I, II, III". 1982, Russian Math. Surveys 37, pp 61-144.
[Po]
Power, S. C. "Hankel operators on Hilbert space". Bull. Lond. Math. Soc. 12, pp. 422-442.
1980,
[Po2 ]
"Hankel operators on Hilbert space". Pitmann Books LTD., London.
1982,
[RT]
Ricci, F. and Taibleson, M. "Boundary values of harmonic functions in mixed norm spaces and their atomic structure". 1983, Ann. Sc. Norm. Super. Pisa, Cl. Sci., IV, Ser. 10, pp. 1-54.
[Rl]
Rochberg, R. "Trace ideal criteria for Hankel operators and connnutators". 1981, Indiana U. Math. J. 31, pp. 9l3-925.
[R2]
"Interpolation by functions in Bergman spaces". 1982, Mich. Math. J. 29, pp. 229-236.
[RS]
Rochberg, R. and Semmes, S. "A decomposition theorem for BMO and applications". 1984, preprint.
DECOMPOSITION THEOREMS FOR BERGMAN SPACES
277
[S]
Sennnes, S. "Trace ideal criteria for Hankel operators and applications to Besov spaces". 1984, Integral Equations and Operator Theory 7, pp. 241-281.
[S2]
"The Cauchy integral and related operators on smooth curves". 1983, Ph.D. thesis, Washington U., St. Louis, MO.
[So]
Soria, F. "Characterization of classes of functions generated by blocks and associated Hardy spaces". 1984, preprint.
[Su]
Sundberg, C. "Interpolation of values by functions in BMO". 1983, preprint.
[TW]
Taibleson, M. and Weiss, G. "Certain function spaces connected with almost everywhere convergence of Fourier series". [BCFJ], pp. 95-113.
[U]
Uchiyama, A. "A constructive proof of the Fefferman-Stein decomposition of BMO (Rn)". 1982, Acta Math. 148, pp. 215-241.
[V]
Volberg, A. "Two remarks concerning the theorem of A. Axler, S.-Y.A. Chang and D. Sarason". J. Operator Theory 7, pp. 209-218.
[W]
Wilson, M.
Manuscript, 1984.
)PERATOR-THEORETIC ASPECTS OF THE NEVANLINNA-PICK INTERPOLATION PROBLEM
Donald Sarason Department of Mathematics Universiry of California Berkeley, CA 94720, U.S.A. The Nevanlinna-Pick problem is the problem of interpolating prescribed values on a given set of points in the unit disk by means of a holomorphic function obeying a prescribed bound. These lectures are intended to bring out certain operator-theoretic aspects of that problem. Two approaches to the problem will be discussed, the original function-theoretic one of R.Nevanlinna and a recent operator-theoretic one due to J. A. Ball and J. W. Helton. The latter approach will be employed to study the extension problem for Hankel operators.
1.
Lecture 1:
Introduction; Preliminaries on Krein Spaces
Lecture 2 :
The Schur Algorithm
Lecture 3:
The Nevanlinna-Pick Problem as a Problem on Extensions of Operators
Lecture 4:
Extensions of Hankel Operators
Lecture 5:
Extensions of Hankel Operators (Continued)
INTRODUCTION; PRELIMINARIES ON KREIN SPACES These talks concern a basic problem in function theory first
studied nearly eighty years ago by C.Caratheodory (7).
The focus
here will be on a variant of Caratheodory's problem introduced by G.Pick (17) and R.Nevanlinna (15) which is conveniently stated in two parts.
Let Hoo denote, as usual, the space of bounded holo279
s. C. Power (ed.), Operators and Function Theory, 279-314. e 1985 by D. Reidel Publishing Company.
280
D. SARASON
morphic functions in the open unit disk, The closed unit ball in Hoo
D,
of the complex plane.
(relative to the supremum norm) will
be denoted by ball Hoo .
(A)
If
zl"",zn
are distinct points of
D and w1, ... ,wn
are
complex numbers, under what conditions does there exist a function (B)
00
in ball H
If such a function
such that
W., J
j = 1, ... ,n ?
exists, can one describe the class
of all such functions?
Nevanlinna, although publishing a few years later than Pick, was unaware of the latter's work, due to the interruption in communications caused by the First World War.
The approaches of
Nevanlinna and Pick are quite different; basically, Pick's method handles problem (A) and Nevanlinna's problem (B).
The differences
between Caratheodory's problem and the Nevanlinna-Pick problem are technical rather than substantial, and it is easy to subsume both problems as special cases of a single one.
(In Caratheodory's
problem, one assigns Taylor coeff.icients at
0
rather than
function values, and one interpolates by holomorphic functions with positive real parts rather than by bounded holomorphic functions.)
Of the other early contributors to this theory, only
I.Schur (20) will be mentioned here.
He invented a technique,
now known as the Schur algorithm, for dealing with Caratheodory's problem.
That technique was later adapted by Nevanlinna.
Although the Nevanlinna-Pick problem does not on the surface appear to have anything to do with operator theory, there is a naturally arising question in that theory which leads to it as a special case (19).
The operator-theoretic connections are in fact
detectable in some of the early work.
When the Nevanlinna-Pick
problem first arose, however, operator theory was in its' infancy, and the time was not ripe for the connections to be explored. They have been explored and developed extensively during the last
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
twenty years.
281
(The earliest contribution along these lines
known to the author, due to B. Sz.Nagy and A. Koranyi, dates back nearly thirty years (23).) Somewhat more recently the Nevanlinna-Pick problem has played a role in certain questions in physics and engineering, notably in control theory.
For these questions, and also for purely mathemat-
ical reasons, one is led to generalize the problem to matrix-valued and operator-valued functions. In these lectures, two approaches to the Nevanlinna-Pick problem will be presented:
the original approach of Nevanlinna,
using the Schur algorithm, in Lecture 2, and a recent operatortheoretic one due to J. A. Ball and J. W. Helton, in Lecture 3. Only the scalar-valued problem will be dealt with here; the matrix-valued and operator-valued generalizations will not be touched (although many of the ideas to be presented apply in that generality).
Also untouched will be the well-known commutant
lifting theorem (22),(18), a theorem in operator theory which encompasses a large number of interpolation problems.
Despite
this lack of scope, it is hoped the lectures will provide some of the flavor and some of the substance of a most interesting interconnection between operator theory and classical analysis. The last two lectures, Lectures 4 and 5, concern an operatortheoretic generalization of the Nevanlinna-Pick problem, the extension problem for Hankel operators, which is closely tied to classical function theory.
The approach used is the one of Ball
and Helton. As far as proofs go, ample but not full details have been provided.
The omitted details are for the most part routine.
There is a vast literature on our subject.
The list of
references here has been kept to a minimum in accord with the limited scope of these lectures. Prerequisites for reading this article are a basic knowledge of operator theory in Hilbert space and of the theory of Hardy spaces in the unit disk.
In regard to Hardy spaces, we shall
282
D. SARASON
need in particular the notions of inner functions and outer functions, and A.Beurling's well-known invariant subspace theorem. There are several excellent references:
(8),(9),(12).
The Ball-
Helton approach relies upon some of the machinery of Krein spaces. The relevant material is discussed in the remainder of this introduction. Krein Spaces A Krein space is a Hilbert space with an additional bit of structure, an indefinite inner product.
One can define a Krein
space as a pair (H,J), where H is a Hilbert space and J symmetry on H,
that is, a self-adjoint unitary operator on H.
To elimina te trivial cases we assume that J identity nor its negative. x
and
y
is a
is neither the
The J-inner product of the two vectors
in H is denoted [x,y]
and defined to be (Jx,y>
(where
denotes the given inner product on H). The spectrum of J
is {l,-l}.
are commonly denoted by H+ and H_,
The corresponding eigenspaces and the orthogonal projections
onto these eigenspaces are denoted by P+ and P . J = P+ - P _,
Thus
and
What follows is a review of the basic notions and facts from the theory of Krein spaces that are needed in the sequel. proofs will be given.
No
The standard references on the subject
are (4),(6),(13). The vector x [x,x] ~ 0
in the Krein space H is called positive if
and negative i f
[x,x] ;;; O.
A subspace of H is called
positive if it consists of positive vectors and negative if it consists of negative vectors.
(Convention:
subspaces of a Hilbert
space are assumed to be closed.) If
T is a contraction operator whose domain is a subspace
of H
and whose range is contained in H ,
of T,
is a positive subspace of H.
+
-
then G(T),
the graph
Conversely, each positive
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
283
subspace of H is the graph of such a contraction, called the angular operator of the subspace.
A positive subspace is maximal
positive, that is, not properly contained in another positive subspace, if and only if the domain of its angular operator is all of H+.
A positive subspace is called uniformly positive if
the norm of its angular operator is less than 1.
(In particular,
the trivial subspace, {O}, is uniformly positive.)
There are
analogous connections between negative subspaces and contractions whose domains are in H
and whose ranges are in H+.
A number of basic Hilbert space notions have J-analogues. For instance, two vectors x and y in H are called J-orthogonal, written x [1] y,
if
[x,y] = O.
The J-complement of the subspace M
of H, written M[1], is the set of vectors in H that are J-orthogonal to every vector in M; it is related to the usual orthogonal complement of M by the equality
M[1] = JM1.
the operator T on H, written T[*], [Tx,y] = [x, T[*]y]; the equality
are assumed to be bounded.) [x,x] for all x,
is defined by the relation
it is related to the usual adjoint of T by
T[*] = JT*J.
adjoint i f T=T[*].
The J-adjoint of
(Convention: Hilbert space operators The operator T is called J-self-
I t is called a J-contraction i f [Tx,Tx];;;
a J-isometry i f [Tx, Tx] = [x,x] for all x,
J-unitary if it is both invertible and a J-isometry. operators are those satisfying
T[*]T = TT[*] = 1.
and
The J-unitary An operator on
H is called a J-projection if it is a J-self-adjoint idempotent. The subspace M of H is called regular if it can be written as the J-orthogonal sum of a uniformly positive subspace, M+, and a uniformly negative subspace, M_.
One can make such a subspace
M into a Krein space in its own right by retaining the J-inner product but introducing a new positive definite inner product. The new inner product is the one that agrees with [ :find with - [ ,
on M.
] on M+
It makes M into a Hilbert space with a
norm equivalent to the one inherited from H.
One consequence is
. that a linear manifold in the regular subspace M is dense in M if and only if no nonzero vector in M is J-orthogonal to it.
It
284
D. SARASON
can be shown that an equivalent condition for the regularity of M is that M be the range of a J-projection; another equivalent condition is the equality
H = M +M[l].
In particular, the regu-
lari ty of M implies that of M[ 1] . The simplest example of a Krein space is the two-dimensional space «::2 with the symmetry and y = y
+
@
y
-
=(~ _~).
J
Thus, for x = x+
@
x_
in C 2 we have (x,y) [x,y]
If
T = (:
~)
is an invertible 2 x 2 matrix, then T corresponds
both to an operator on C 2 and to the linear fractional transforma-----'"" az+b One easily tion z --,. cz + d on C, which we also denote by T. verifies that the linear fractional transformation T maps the unit disk into itself if and only if the operator T transforms negative vectors to negative vectors.
Further, one can show that the
preceding condition holds i f and only i f T is a scalar multiple of a J-contraction.
Similarly, the linear fractional transforma-
tion T maps the unit disk onto itself if and only if the operator T is a scalar multiple of a J-unitary operator. is J-uni tary i f and only i f ab=cd.
=
Id I 2 - Ib 12
I t is a J-contraction i f and only i f
Id 12 - Ib I2
;;; 1,
= 1
and
lal2-lcl2 < 1,
and
Iab 2.
Ia 12 - IC 12
The operator T
cd 12
~
(1 -
Ia I 2 + I c I 2)( Id I2 - Ib I 2 -
1) .
THE SCHUR ALGORITHM The Schur algorithm effectively handles part (B) of the
Nevanlinna-Pick problem.
It is based on the Schwarz-Pick lemma,
which is the conformally invariant form of the well-known Schwarz lemma.
For
a
in D we let b a
factor for D vanishing at
a:
denote the normalized Blaschke
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
~(a-_z)
b (z) a (b o(z) = z).
1-
az
This is the unique conformal automorphism of D that
to 0 and 0
sends a
285
derivative at 0,
to the positive real axis (or has a positive
in the case a = 0) •
Schwarz-Pick Lemma (9, p. 2) :
If the function cp is in
ball HOD and is not a unimodular constant, then, for any
a
in D,
The preceding inequali ty is strict at all points of D other than a unless
cp
is a conformal automorphism of D.
Turning to the Nevanlinna-Pick problem, we assume given n distinct points zl' ... ,zn in D and n
values wI' ... ,wn which
are to be interpolated along zl' ..• ' zn by a function in ball Hoo • We assume the interpolation is possible, and we ask for a description of the most general function performing it. we shall write bj
for b zj .
Suppose the function cp in ball Hoo satisfies j = 1, ... ,no
cases:
We look first at the point zl
IwI I = 1
and
IwI I < 1.
the maximum principle that cP in particular, w2
For simplicity
' •••• wn
CP(Zj) = Wj ,
and distinguish two
In the former case we conclude from is identically equal to wI
all coincide with wI).
(and so,
In the latter
case, the Schwarz-Pick lemma tells us that the function
(4)-:1 )
bl
l-w l 4>
1
is in ball HOD.
'1'
Solving the preceding equation for
we find that b 1
wb 1
1
+ 1
4>
in terms of
286
D. SARASON
This we can re-express as
~
'I'
-- U 1 '1'1 ~
where U1
'
is the 2 x 2
matrix function
(By U1
is meant the function whose value at
the image of
z
is
U1(z)
under the linear fractional transformation
induced by the matrix U 1(z) .)
The factor
(1 -
2 -~ Iw 1 I)
definition of U1 was inserted because it makes U1
in the
a J-unitary
matrix at each point of the unit circle (relative to the symmetry J =
(~ _~).
The matrix U1
the unit disk, and If
1\
is J-contractive at each point of
det U1 = b 1 •
is any function in ball Hoo
ball Hoo, and its value at
zl
is wI'
the function U1¢1
is in
The function U1
will
satisfy the remaining interpolation conditions if and only if
precisely, one needs
--=--.-1 b 1 (z.) J
(Wj - wI 1
)
j=2,oo.,n
w w - 1 j
The procedure above thus reduces the original n-point interpolation problem to an (n-l)-point problem with revised data. We can now apply the same procedure to the (n-l)-point problem. We find that either
is a unimodular constant -- in which case cp
must be a Blaschke product of order 1 -- or else we reduce the (n-l)-point problem to an (n-2)-point problem.
In the latter case,
we can apply the same procedure again, and so on. There are two possibilities:
either the procedure terminates
at the jth stage (j = 0, .•. ,n-l), or else i t can be carried through the nth stage. product of order
In the former case
j.
nate, the general
is unique and is a Blaschke
In the latter case the problem is indetermihaving the form u1jJ with 1jJ
where U is a certain 2 x 2 matrix function. U = U 1 U2
•••
Un'
where
00
in ball H ,
Precisely,
287
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
u.J
,
Wj
being the datum going with the point Zj
interpolation problem at stage j.
for the revised
The matrix function U is
J-unitary on the unit circle and J-contractive in the unit disk (since each of its factors is), and
det U
= b l b 2 •.. b n = b,
the Blaschke product for the points zl, ... ,zn. We thus see that, in case our interpolation problem is solvable but indeterminate, the family of solutions is the image of ball Hoo under a "linear fractional" map. that case, we write
U
(~ ~).
=
Assuming we are in
One easily verifies that the
entries p,q,r,s are rational functions of order at most n, poles confined to the set J-contraction in D, J-unitary on addition Because
aD
ps and b
one has
one has
=
ps - qr
=
=
Because U is a
1, •.. ,n}
I s I ;;; 1 there.
Ipi = I s I and
Isl2 -lrl2 = 1. det U
{zjl: j
Because U is
I q I = I r I there, and in
In particular, Ipsl> Iqrl b,
with
on
aD.
it follows by Rouche' s theorem that
have the same number of zeros in D, so
p has n
zeros
in D and is thus precisely of order n. In addition, one has the relations r(z)
=
p(z)
----l-
b(z)s(z
=
),
b(z)q(z-l)
To verify these we note that, by the J-unitarity of U on aD, one has there the equality U- l in other words, 1 ( s b -r
-q) p
=
(~ -~) -q
s
which says that the desired equalities hold on
aD.
By analytic
continuation, the equalities hold everywhere. The procedure above, the so-called Schur algorithm, completely answers part (B) of the Nevanlinna-Pick problem.
What about part
(A)?: gi ven the values w1 ' ••• ,wn ' how can one recognize whether " there is a function in ball H~ that interpolates those values
288
D. SARASON
along Z1 ' .•. , zn? this, of sorts.
The Schur algori thm also gives an answer to One can program each step of the procedure
entirely in terms of the data Zl ' ••
Zn' WI'··· ,wn • Given that data one can attempt to carry out the procedure. The interpolation • ,
problem will have a solution if and only if the procedure does not break down but can be brought to a successful conclusion, either after a full complement of n steps or at some intermediate stage. This answer to (A) is very implicit.
The operator-theoretic
approach, to be discussed in the next lecture, offers a more explicit answer. We take up now the infinite Nevanlinna-Pick problem. 00
(zn) 1 and
Let
be an infinite Blaschke sequence in D, without repetitions, 00
(wn ) 1 a sequence of complex numbers. We ask whether there is 00 00 a function in ball R which interpolates the sequence (wn)l along 00
the sequence (zn)l' and, in case there is, we ask for a description of all such functions.
The focus here will be on the latter ques-
tion, which was treated definitively by Nevanlinna in 1929 (16). A sketch of Nevanlinna's analysis follows. The idea is to approach the infinite interpolation problem as the limit of its finite sections.
For each positive integer n
we consider the finite Nevanlinna-Pick problem with data zl, ..• ,zn' w1, ..• ,wn . We restrict our attention to the most interesting case, that where each of the finite problems is solvable but indeterminate.
In that case, as we saw above, the general solution of the
problem with data zl' ••. ,Zn' wI' .•. ,wn has the form un 1/! with 1/! in ball Roo and Un a certain 2 x 2 matrix of rational functions which is J-unitary on 3D and J-contractive in D, and whose determinant is bn ,
the Blaschke
(Note the change in notation.)
product with zeros Z l' •.. , Zn • We write
For Un it is convenient to take not the matrix function constructed above but rather a normalized version with the property rn (0) = O.
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
289
One can achieve the normalization by mUltiplying the originally constructed matrix function from the right by a suitable constant J-unitary matrix. From the J-contractiveness of Un Isnl ~ 1 there.
aD
On
one has
the J-unitarity of Un.
in D it follows that
Isnl = IPnl ~ Iqnl = Irnl,
Consequently, the functions l/sn'
due to Pn/sn'
00
qn/sn'
rn/sn all belong to ball H .
For z
in D the set
disk whose radius,
Pn (z),
{Un (z)ljJ(z): IjJ E ball Hoo }
is a closed
is given by
Pn(z) This radius is obviously nonincreasing with respect to n,
and a
normal families argument shows that the infinite interpolation problem is indeterminate if and only if at least for one point z for
Pn (z)
in D.
lim Pn(z)
n-+ OO
The denominator in the expression
is bounded from above by
I sn (z) 12
and from below by
(1-lzI2)lsn(z)12 (by Schwarz's lemma, since rn/sn and vanishes at 0).
is positive
is in ball Hoo
We thus have the inequalities
Since b n converges to the Blaschke product for the sequence (zn): (which we denote by b), and since the boundedness of the functions s
n
at one point of D implies their uniform boundedness
on compacta in D (by Harnack's inequality, applied to the functions log I sn I)' we infer from the inequalities above that the infinite interpolation problem is indeterminate if and only if the functions
sn are uniforthly bounded on compacta in D.
Suppose the preceding condition holds. maj orizes IPn I, I qn I and 00
00
00
We know that Isnl
I rn I in D, so each of the sequences 00
(Pn) l' (qn) l' (rn ) l' (sn) 1 is uniformly bounded on compacta in D. One can thus select a subsequence (nj)7 of the positive integers such that the corresponding subsequences of (s)oo n
1
(p )00, (q )00, (r )00, nl nl nl all converge uniformly on compacta in D, say to the func-
290
D. SARASON
tions p,q,r,s,
respectively.
Let U be the matrix function
(~ ~).
It is then a simple matter to show that the functions oo in ball H solving the infinite interpolation problem are the functions UIj! with Ij! in ba11 Hoo . The matrix U(z),
being the limit of J-contractive matrices,
is J-contractive for each z lim det Un'
n~oo
in other words,
in D.
The determinant of U is
lim b n , n~oo
which is b.
The entries of
U are in the Nevanlinna class (since lis, pis, q/s, rls, being limits of functions in ball Hoo , are in ball Hoo ). Hence U has a
aD
boundary function, defined almost everywhere on
by means of
radial limits, and the boundary function is clearly J-contractive almost everywhere. The question now arises whether the boundary function of U
aD.
is not actually J-unitary almost everywhere on
To show that
a J-contractive 2 x 2 matrix is J-unitary, it suffices to show that the linear fractional transformation it induces sends
aD
aD.
to
Using this observation, one can reduce the task of proving U is J-unitary almost everywhere on
aD
to the task of proving that,
for every constant unimodular function
A,
the function UA is an
inner function, a task which is accomplished by the following ingenious argument of Nevanlirtna. A be as above, and let cp = UA.
Let
The function cp
solves
each of the finite interpolation problems, so for each positive integer n -1
Un
Ij!n
it can be written as Unlj!n with Ij!n
in ball Hoo .
exists in D except on the zero set of b n , =
U~IUA
except on that set.
on the zero set of b,
As
U~;
we can write
~ U- 1
we can conclude that
in D except
Ij!n. -> A off that J
set, and hence actually throughout D.
The strategy now is to
show that the preceding limit relation would fail if an inner function. On
aD
in other words
we can write
pn cj>- r n -qnCP + sn
II
cP - (qn/sn)
=
As
1- (qn/sn)cj>
cp were not
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
I
(since
Iqnl2
Isn12-1
=
where Isnl
aD
(1-lsnl-2)~
1
1 + I qn I sn I I
E(c) be the subset of
291
1+
on aD).
(1-lsnl-2)~I
For O
where I
let
F(n,a) the subset of aD
By the preceding inequality, on
E(c) n F(n,a)
we have C
+
(1 - a
1 + (1- a
Because ljJn (0)
-2 ~
)
K(c,a)
-2 k
)
2 C
is the average of ljJn over
aD,
the last inequality
implies IljJn (0) I
ii
1- (1- K(c,a»
meas(E(c) n F(n,a»
(where "meas" here refers to normalized Lebesgue measure). number K(c,a)
The
IljJn.(O) I -> I A I
is less than 1, so, because
J
it follows that, for fixed meas(E(c) n F(nj,a»
Our aim is to show that
a
and ~
=
1,
c, (j-+oo)
0
meas E(c)
=
O.
We shall show that
the preceding limit relation is impossible unless that is the case. We need an estimate of the sizes of the sets F(n,a), which we obtain from the observation that log I sn lover that
aD.
Letting
M = sgP log I sn (0) I ,
1- meas F(n,a) ii M/10g a,
large, we can guarantee that
log Isn(O) I is the average of
so, by choosing
meas F(n,a)
1- meas F(n,a) < ~ meas E(c) meas(E(c) n F(n,a»
for every n,
> ~ meas E(c)
a
sufficiently
is closer to
any preassigned positive number, for every n. meas E(c) were positive, we could choose
we conclude
a
1
than
But then, if
large enough to make and we would have
for every n,
in contradiction
to what we established in the preceding paragraph.
The proof that
292
D. SARASON
U is J-unitary almost everywhere on
aD
is now complete.
Two concluding remarks: We defined the matrix function U by taking the limit of
1.
00
a convergent subsequence of (Un )I' converges to U.
Actually the original sequence
The proof of this depends on a uniqueness
result for U.
2.
We insisted in the discussion above that the points zn
be distinct, but that was solely to simplify the exposition. The Nevanlinna-Pick problem with multiple nodes can be handled by the same methods. 3.
THE NEVANLINNA PICK PROBLEM AS A PROBLEM ON EXTENSIONS OF OPERATORS We now give an operator-theoretic reformulation of the Nevan-
linna-Pick problem and analyze the problem using operator-theoretic techniques.
This approach not only recaptures from a completely
different viewpoint the results obtained in the last lecture by means of the Schur algorithm, it also leads to a simple necessary and sufficient condition for the interpolation problem to have a solution. As before, we consider distinct points zl"" ,zn in D and complex numbers wI"" ,wn which we wish to interpolate along 00 Zl' ... ,zn by a function in ball H. We let b denote the Blaschke product for the sequence zl ••.• 'zn. In the Hilbert space H2 of square-summable power series in the unit disk, we form the subspace space of dimension n, j = 1,0" ,no
=
H2
e
spanned by the functions
The function k j
Zj: (h,kj> = h(zj)
Mb
bH2;
k j (z)
=
(1- zjz)
-1
,
is the kernel function for the point
for h in H2.
The shift operator on H2 will be denoted by S: zh(z).
it is a sub-
(Sh) (z) =
The subspace Mb is S*-invariant, being the orthogonal
complement of the S-invariant subspace bH 2
* k j are eigenvectors of S:
In fact, the functions
S*k j = -zjKjo
On Mb we define an operator A by setting
The
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
operator A*
293
obviously commutes with S* 1Mb.
We ask whether A*
can be extended to an operator on H2 which commutes with S*. An operator on H2
which commutes with S*
an operator which commutes with S,
is the adjoint of
and, as is well known and
easily shown, the operators commuting with S are the multiplica00 tion operators induced by the functions in H. Suppose cJ> is a function in Hoo ,
the adjoint of whose induced multiplication
operator extends A*. (k i , A*k j ) equals
Then
cJ>(Zj) = Wj
for each j,
because
Wj (k i , k j ) on the one hand, while on the other
hand i t equals (<j>k i , k j ), and (<j>k.,k.) 1
J
The same argument shows that, conversely, if cJ> oo
H
such that
cJ>(Zj) = Wj
is a function in
for each j, then the adjoint of the
multiplication operator induced by cJ> extends A*. norm of the multiplication operator on H2
Moreover, the
induced by cJ> is 1IcJ>1100.
We see therefore that part (A) of the Nevanlinna-Pick problem can be reformulated thus:
Under what conditions can A*
to an operator on H2
of norm at most
be extended
1 which commutes with S*?
An obvious necessary condition for the existence of the
desired extension of A*
(and hence for the solvability of the
associated interpolation problem) is the inequality If
is a typical vector in Mb ,
I i,j and
then
cic. (k., k. ) J 1 J
.
I
i,j The inequality
IIA*II ~ 1 .
wiCiWjC j (ki,kj )
IIA*II,;;; 1
definiteness of the matrix
is thus equivalent to the positive semi-
294
D. SARASON
the so-called Pick matrix associated with our interpolation problem.
That the positive semidefiniteness of this matrix is in
fact equivalent to the solvability of the interpolation problem goes back essentially to Pick in his original paper on the subject. Recently J. A. Ball and J. W. Helton (5) had the very nice idea of putting the preceding extension problem into a Krein space context. That device replaces the operator extension problem with a subspace extension problem.
Here is a sketch of the reasoning.
We deal first with the case space
H
= H2
@ H2,
the symmetry
J
=
II A*II < 1.
We form the Hilbert
which we regard as a Krein space relative to
(~ _~).
The shift on H is the operator S @S,
but we shall be sloppy and for convenience denote it simply by S. It is a J-isometry as well as an isometry, and The graph of A*, h E~},
that is, the subspace
=
S[*] G(A*)
S*.
{h @ A*h:
=
is S*-invariant because A* commutes with s*IMb'
also uniformly positive because of our assumption that
It is
IIA* II < 1.
The problem of finding an extension of A* which commutes with S* and has norm at most 1 is equivalent to the problem of finding an S*-invariant subspace of H which contains G (A*) maximal positive.
and is
To establish the existence of such a subspace
we analyze the subspace
N
= G(A*)[l]
The subspace N is S-invariant.
We apply to it the J-analog ue
of a method used by P. R. Halmos (10) to analyze shift-invariant subspaces.
Because N is the J-complement of a regular subspace
it is regular. IIA *11 < 1.) regular.
(It was to obtain this property that we assumed
Since S is a J-isometry the subspace SN is also Therefore N is the vector sum of the two mutually
J-orthogonal, regular subspaces SN and SN is the vector sum of S2N and SL, so
L
=
N () (SN) [1].
N = L+ SL+ S2N,
three subspaces on the right being mutually J-orthogonal.
But then all Iterat-
ing this reasoning we find that, for any positive integer n, N
L + SL + .. ,
all subspaces on the right being mutually J-orthogonal.
This
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
295
implies (because L and its images under powers of
S are regular)
that any vector in N which is J-orthogonal to SnL for all n ~ 0
n SnN and so mus t be O. Because of the regulari ty n>O 2 it follows that N is spanned by the subspaces L, SL, S L,
mus t be in of N,
The subspace L is neither positive nor negative.
In fact,
one easily sees that if L had either of these properties, then N would have the same property. because its J-complement,
But obviously N is not positive,
G(A*),
is positive.
And if N were
negative then its J-complement, being positive, would have to be maximal positive, which it obviously is not. Since L is neither positive nor negative, it contains a pair of vectors x I ,x 2 such that
[xI,x l ]
=
I
=
-[x 2 ,x 2 ]
and
2
Let N+ be the subspace spanned by xl' SX I ' S Xl'
[x I ,x 2 ]
=
O.
One easily
checks that N+ is a positive subspace, so the subspace G(A*) + N+, being the J-orthogonal sum of two positive subspaces, is itself positive.
It is also S*-invariant, because S*xI
is in G(A*)
(being J-orthogonal to N). We shall show that the subspace positive.
G (A*) + N+
is maximal
For that we need only to show that the image of
G(A*) +N+ under
P+ (the orthogonal projection operator from H
to its first summand) is all of H2.
Since that image is closed
and obviously contains Mb' we need only to show that it contains Snb for n~ O. Now because S * xl lies in G(A*), the components are sent by S*
in Mb + Cb.
into Mb , which means these components lie Hence the image under P+ of G(A*) + CX I is contained
in ~ + Cb.
Because P+
of xl
of G(A*)
is one-to-one on
under P+ has codimension I
The former image is Mb , the image of
G (A*) + N+,
the image
in the image of G(A*) + Cx l .
so the latter one must be Mb
G(A*) +N+ under P+ contains
b.
+ Cb.
Hence
The obvious
iteration of this reasoning shows that the image of G(A*) + N+ under P+ contains Snb
for n=I,2, ••. ,
maximal positivity of G(A*) + N+.
thus establishing the
296
D. SARASON
We have now produced a maximal-positive S*-invariant subspace of H containing G (A*) ,
so we have shown that A* has an extension
of the desired kind and thus that the condition
IIA*" < 1
is
sufficient for the solvability of our interpolation problem. The sufficiency of the weaker condition "A*" ;:;; 1 follows now by a simple limit argument.
We have thus responded to part (A) of
the Nevan1inna-Pick problem. We proceed to analyze the situation in greater detail in order to answer part (B) from an operator-theoretic viewpoint, maintaining the assumption "A*" < 1.
The space H2,
originally introduced
as a space of ho10morphic functions in D, will be identified in the usual way with the corresponding space of boundary functions on
aD
(a subspace of the L 2 space of normalized Lebesgue measure). In this way H becomes identified with a subspace of the C 2-va1ued L2 space of normalized Lebesgue measure on aD. We think of C2 as a Krein space in the way mentioned at the end of Lecture 1 and used in Lecture 2. First we observe that the equalities and
aD,
[xl ,x 2 ] = 0 also hold pointwise on
[xI(ei8),xI(ei8)] = 1 = -[x2(ei8),x2(ei8)] for all e i8 in as n
aD.
[xI,x l ] = 1 = -[x 2 ,x 2 ] in other words, and [xI(ei8),x2(ei8)]=O
In fact, since [xl,Sn xl ] is 1 or 0 according
is 0 or positive, and since 1f
1
21f
f
-1f
we see that the function [xl(e
[xl (e
i8
i8
), xl(e
) 'Xl (e i8
)]
i8
)]
e- in8 d8
has the same Fourier
series as the constant function 1, and thus it equals the constant function 1, at least almost everywhere. the components of xl
lie in Mb + Cb
Moreover, we know that
and so are rational func-
tions. In particular, they are continuous, so actually "8"8 i8 [Xl (e 1 ) 'Xl (e 1 )] = 1 for all e on aD. The other two relations are established by similar reasoning. We can now see that the dimension of. L is exactly 2.
In
fact, if x is a vector in L which is J-orthogona1 to both xl
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
then by the reasoning above x(e ie )
and x 2 ' both xl
297
·e )
(e 1
is J-orthogonal to
which means x = 0 because, by the result in the last paragraph, xl(e ie ) and x 2 (e 1·e ) form for each e a basis for «:2. Thus the vectors xl and x 2 form a basis for
and
·e )
x/e 1
L.
We write matrix function of xl
xl
p Ell rand
(~ ~).
x 2 = q Ell s,
and we let U denote the
It was noted above that the components
are rational functions, and the same is clearly true of the
components of x 2 . in H 2;
for all e,
The entries of U are thus rational functions
in particular, they lie in Hoo
J-unitary at each point of the relations
aD
The matrix function U is
due to the pointwise validity of
[xl'x l ]=1=-[x 2 ,x 2 ],
[x l ,x 2 ]=O.
It follows that
U acts via multiplication as a bounded operator and as a J-isometry from H to H.
Also, the range of U is closed, because U
acts via multiplication as an invertible operator from
L2 Ell L2
onto itself (its inverse being its pointwise J-adjoint). is clear that UH is contained in N and contains for all nonnegative integers n,
snxI
As it and snx2
we can conclude that UH = N .
We are now in a position to rederive the description, obtained in Lecture 2 by means of the Schur algorithm, of the most general solution of our interpolation problem.
In operator-
theoretic terms, we want to describe the most general operator on H2
that commutes with S,
adjoint extends A*.
has norm at most 1, and whose
Equivalently, we want to describe the most
general maximal-negative S-invariant subspace of H contained in N. If N'
is such a subspace then, by virtue of the properties of U -1
mentioned in the last paragraph, the subspace U
N
,
is maximal
negative; it is also obviously S-invariant (since U commutes with S).
Hence, every subspace of the kind we want is the image
under U of a maximal-negative S-invariant subspace of H. The converse of the last statement is also true but it takes a bit of work to establish it. S-invariant subspace of H,
If N"
is a maximal-negative
then UN" is clearly negative and
298
D. SARASON
S-invariant; it is also clearly maximal negative in N, clearly maximal negative in H. is uniformly negative.
but not so
The situation is simplest when Nil
Then H is the J-orthogona1 sum of Nil
and
the positive subspace
N'" = (N") [1] ,
sum of UN"
Consequently H is the J-orthogona1 sum of
and UN'" •
the negative subspace UN"
so N is the J-orthogona1
and the positive subspace
G (A*) + UN'" ,
which implies the maximal negativity of UN". Applying the preceding observation to the special case where N" -- {a} m w H2,
the vector x 2 under P
. we see t h at t h e S' -1nvar1ant sub space generate d b y
is maximal negative.
The image of that subspace
(the orthogonal projection operator from H to its
second summand) is the S-invariant subspace generated by s (the second component of x ), and hence
s
must be an outer
function. Consider now any maximal-negative S-invariant subspace N" of H.
Such a subspace is the graph of the multiplication operator in ball Hoo :
induced by a function tjJ thus have
N"
{tjJh GO h: h E H2}.
UN" = {(ptjJ+ q)h ~ (rtjJ+ s)h: h E H2}.
maximal nega ti ve amounts to showing that tion.
We do that by writing
outer, and so is sum of
1
1 + tjJr / s
is an outer func-
rtjJ+s=s(l+tjJr/s).
because tjJr / s
is
We know s i s
lies in ball Hoo (and the
and a function in ball Hoo is always outer (12, p .117) .
Hence rtjJ + s
is outer, being the product of two outer functions,
and the maximal negativity of UN" see that UN" by UtjJ.
rtjJ + s
To show UN"
We
is established.
Moreover, we
is the graph of the multiplication operator induced
We can conclude that the general solution of our interpo00
1ation problem has the form UtjJ with tjJ
in ball H
At this point we have recaptured the essence of what was established in the last lecture by means of the Schur algorithm. The analysis can be further refined; we mention a few facts but do not pursue them in detail. been established about
First of all, from what has already
U it is not too hard to show that its
determinant 1S a constant multiple of b,
so, rep1acinl!, xl'
by a constant mUltiple of itself, we can assume
det U
=
b.
say, One
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
then obtains the equali ties last lecture. xl
and
Second,
299
p = bs and
r = bq on
aD
as in the
U obviously depends upon the choice of
x 2 ' but only to within multiplication from the right by a
constant J-unitary matrix, and that in fact is the extent of the arbitrariness of U.
In particular, a matrix function U constructed
by the present methods and one constructed by the Schur algorithm for the same interpolation problem are the same to within a constant J-unitary factor on the right. Third, it is natural to ask if one can somehow express the entries of the matrix function U in terms of the operator A. We shall go into that in a little detail. concerned wi th q
and s,
We need only to be
the componen ts of x 2' because, as
mentioned in the preceding paragraph,
p
and
r
can be expressed
in terms of them. The subspace N is easily seen to be the orthogonal sum of bH 2 E9 bH 2 and
{Af E9 f: f
Mb },
E
the graph of A.
We have noted
+ Cb. Conse(Af @ f) + (ab E9 Sb)
that the components of a vector in L belong to Mb quently, any vector x for some f
in Mb
in L has the form
and scalars a, S.
x =
That x
is in L means that
S*x belongs to G (A*), in other words, that (S*Af E9 S*f) for some g
in Mb.
S*Af
+
+ (as*b E9 SS*b)
g E9 A*g
Equating components, one finds
as*b
S*f
g ,
+ SS*b
A*g
which combine to give S*f
+ SS*b
Thus the function f
S*A*Af
vector
aA*S*b
and scalars a, S must satisfy the equality S*(l- A*A)f
for x
+
to belong to L. x = (Af E9 f)
+
(aA* - S)S*b
The reasoning is reversible, so the (ab E9 I3b)
will belong to L i f f, a, 13
300
D" SARASON
satisfy the equation above.
The object now is to determine a
simple solution of that equation which will yield a strictly negative vector x in L and thus, after normalization, one possible choice of x 2 ,
explicitly expressed.
It turns out one
arrives at such a solution starting from the premise ex = O.
The
reasoning is slightly different for the cases b (0) of 0 and b (0)
= 0,
corresponding, respectively, to the invertibility and noninvertibility of S* 1Mb. function s
We omit further details.
c[(l-A*A)-lu + b(O)b]
S
where
The particular
one obtains is
u = l-b(O)b
(the projection of the constant function 1
on Mb ) and
The corresponding function
q
is
cA(l - A*A)-l u.
Final"ly, a few words are in order concerning the case IIAII
=
1. oo
In that case we know there is at least one function
This
cP
is in
To see that is
so one takes a nonzero function f
in Mb such that IIAfll2 = IIf1l 2 . It is then a simple matter to show that
desired conclusions follow.
Details can be found, for example,
in (19). 4.
EXTENSIONS OF HANKEL OPERATORS The operator-theoretic reformulation of the finite Nevanlinna-
Pick problem given in the last lecture applies equally to the infinite Nevanlinna-Pick problem; one merely has to replace the finite Blaschke product b by an infinite Blaschke product. subspace
~
The
becomes infinite dimensional, and certain technical
difficulties arise, but the Ball-Helton approach remains effective. In fact, one can replace
b
by an arbitrary inner function,
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
301
thereby obtaining the operator-theoretic generalization of the Nevan1inna-Pick problem alluded to in the first lecture.
That
problem, in turn, is subsumed by the problem of extending Hankel operators, the subject of this lecture and the next one. The Hankel operator problem will be introduced from the outset in a Krein space framework. H
=
L2 @ H2,
synunetry
J =
We form the Hilbert space
which we regard as a Krein space relative to the
(~
-f).
(Here L 2 stands, as usual, for the L 2
space of normalized Lebesgue measure on
an,
identified as usual with a subspace of L2.) on H is the operator S defined by
and H2 will be The shift operator
(Sx) (z) = zx(z);
it is the
direct sum of the bilateral shift on L 2 and the unilateral shift on H2. By a Hankel operator we shall mean an operator from {O}@H 2 to (H2)1 @ {a} whose adjoint commutes with S* (more accurately, whose adjoint intertwines s*I{O} @ H2).
s*1 (H2)1 @ {a}
and
If b is an inner function and Ao is an operator on the space ~ = H2 e bH 2 whose adjoint commutes with s*I~, then there is a Hankel operator A naturally associated with Ao' define A one extends Ao
To to all of H2 by making it 0 on bH 2
and then one sets
A(O@h) = bA h @ 0 (h E H2). That A so o defined is actually a Hankel operator is easily verified. The results we shall obtain about Hankel operators thus specialize to results about operators such as Ao -- in particular, to results about Nevan1inna-Pick interpolation. 00 Each function cp in L induces a Hankel operator, denoted here by Acp' of which cp is called a symbol.
The operator Acp'
by definition, maps the vector Oeh to the projection onto (H2)1 @ {a} of CPh @ O. The operator Acp remains unchanged if 00 we modify cp by adding to i t an H function, so Acp actually depends only on the conjugate-analytic part of cp (the projection of cp onto (H 2 )1). The inequality IIAcpli ~ 11>1100 is obvious. In fact, by virtue of the preceding remark, we have IIA>II ~ dist(CP,Hoo) . That the last inequality is actually an
302
D.SARASON
equality is a well-known theorem of Z. Nehari (14) which will come out of the analysis that follows. Given a Hankel operator A, write it as A> with
11 >1100 ;:i! I?
under what conditions can one If A can be so written, can one
describe the corresponding class of functions >?
We shall refer
to this two-part question as the extension problem for Hankel In case A arises from a Nevanlinna-Pick interpolation
operators.
problem in the manner described above, it is merely a mild rephrasing of the operator-theoretic formulation of that interpolation problem discussed in the last lecture. Some explanation is needed for the use of the term "extension problem" for the problem of finding a symbol of prescribed norm for a Hankel operator. tor from
If > is in Loo ,
{oJ ~ H2 to L2 ~ {oJ
the multiplication opera-
induced by > is, in the terminology
of P. R. Halmos (11, p.120), a dilation of the operator A>.
The
norm of that multiplication operator is 11>11 00 •
Such multiplication operators are precisely the operators from {oJ ~H2 to L2 Ell {OJ that commute with S.
Thus, given the Hankel operator A,
the
problem of writing A as A> is the same as the problem of dilating A to an operator from
{oJ Q1 H2 to
L2 ~ {oJ
The problem of finding such a > with
which commutes with S.
11>11 00 ;;;; 1
is the same as
the problem of finding such a dilation of A of norm at most 1. Perhaps the term "dilation problem" would have been more appropriate than "extension problem", although the adjoint of such a dilation of A is a bonafide extension (not merely a dilation) of A*. We now study the extension problem for Hankel operators using the Ball-Helton approach.
The condition IJAII
~
1 is
obviously necessary for the Hankel operator A to be writable as A> with
II C/l 1100 ;;;; 1.
We shall eventually show that the
condition is also sufficient, which is basically the theorem of Nehari already mentioned.
For the rest of this lecture we assume
given a Hankel operator A satisfying IIAII < 1. will be dealt with in the next lecture.)
(The case IIAII = 1
The plan of attack is
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
303
the same as in the last lecture. G(A*) (= {h ~ A*h: hE (H2)1}),
the
which is uniformly positive and S*-invariant.
The
We form the subspace graph of A*, subspace
N = G(A*)[l]
is thus regular and S-invariant.
Our task
is to show that there is a maximal-negative S-invariant subspace contained in N,
and then to classify all such subspaces.
Such
subspaces are the graphs of the multiplication operators whose inducing functions are the symbols for A having Loo -norms of at most 1. L = N n (SN)[l] .
As in the last lecture, we form the subspace
I t follows, as before, that any vector in N which is J-orthogonal
to
SnL
so
n SnN
n~O
must lie in
1
{a}.
n Sn N . 1
is a reducing subspace of
It is also J-orthogonal to G (A *) (H2)1 EP
co
00
for all
00
S
Now
n SnH =L 2 EP{O}, I
contained in L 2 EP{O}.
and hence orthogonal to
Since the latter subspace is contained in no proper
S-reducing subspace of L 2 ~ {O}, it follows that
00
n SnN I
is
trivial, and thus only the zero vector in N is J-orthogonal to SnL for every n~ O.
Since N is regular, we can conclude that
it is spanned by the subspaces L,
SL,
S2 L , . . . .
One proves just as in the last lecture that L is neither positive nor negative and that L has dimension 2. we choose a basis XI ,x 2 [xl
,X 2 ]
= O.
vectors xl'
As before,
[xl 'Xl] = I = -[x 2 'x 2 ]
for L with
Let M+ be the subspace spanned by G(A*) SX I
'
2
S Xl'
and
and the
Reasoning exactly as in the last
lecture, we can show that M+ is S*-invariant and maximal positive. The subspace N_ = Ml 1 ] is thus S-invariant and maximal negative, 00
so we have proved that A has a symbol of L -norm at most 1. (The same conclusion under the weaker assumption IIAII
~
I
follows
now via a standard compactness argument, so we have essentially established Nehari's theorem.
More details are in the next
lecture.) As in the last lecture, we write and we form the matrix function
U =
xl = pEP rand
(~ ~).
x 2 = q ffi s,
The reasoning in
the last lecture shows that U is J-unitary almost everywhere on
304
aD.
D. SARASON
However, there is an added complication now because the
entries of U presumably no longer need to be bounded functions, and if they are not, operator.
U will not act as a bounded multiplication
Nevertheless, as we shall see,
U does give a correspon-
dence between the maximal-negative S-invariant subspaces of HZ
G)
HZ
and the maximal-negative S-invariant subspaces of N,
to the extent it can. Our first step is to show that the function function.
The S-invariant subspace N_
s
is an outer
is irreducible, being
contained in the irreducible S-invariant subspace N, so the subspace
N
e
SN
has dimension 1,
is nontrivial.
The latter subspace in fact
as one can show by the same kind of reasoning
as was used in the last lecture to pin down the dimension of L. Let x3
be a nonzero vector in
N_ = {hx 3: h E HZ}. component of x3 lies in N_,
x" = hox3 in G(A*)
Since N
and hI
That
;
we then have
is maximal nega ti ve, the second
X
z = hohlx3'
under hI (S) *).
s,
Now the vector where ho
S*x"
The vector
hI (S)*s*x z
Since xl
and
it must be that
the second component of
x"
xz '
Xz
is an outer
In fact it is in L because
(due to the equality
x" is J-orthogonal to xl' of x z '
SN
is an inner function.
is also in N.
iance of G (A*)
e
must be an outer function.
so we can write
function in· HZ
N
Xz
S*x"
is
and the invarspan L while
is a scalar multiple is an outer function
is now clear. We are prepared to show that U associates with each maximalnegative S-invariant subspace of HZ @H z
a maximal-negative S-
invariant subspace of H which is contained in N and whose angular operator is therefore a multiplication operator induced Let Nil be a maximal-negative S-invariant A. subspace of HZ@H z , say Nil = {1jJh G) h: hE HZ}, where 1jJ is in
by a symbol for
ball Hoo •
The matrix function U, acting as a multiplication Z , an d one eaSl'1 y c h ec k S LZIT\L operator, sen d s L001T\ w L0 0 .lnto ~ that it sends
Hoo
G)
the linear manifold
Hoo
into N.
Therefore, the closure in H of
U(N"n (HOOQ} Hoo » (= {
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
305
is a subspace of N; we denote it by N'.
It is S-invariant
because U commutes with S and negative due to the J-unitarity of U on
aD.
under P
is an S-invariant subspace of H2
It is in fact maximal negative, because its image containing the func-
tion r1jJ+ s, which is outer (by the same reasoning as was used to deduce a similar conclusion in the last lecture). operator of N'
is the multiplication operator induced by
(p1jJ+q)/(r1jJ+s), for A in ball L wi th 1jJ in ball
The angular
in other words, by U1jJ. 00
The class of symbols
thus consists of at least all the functions
U1jJ
Hoo •
Before proving that the procedure which produces N' from Nil is reversible, in other words, that the class of symbols for A in ball L
00
00
consists precisely of the functions
U1jJ with 1jJ in ball H ,
we deduce an additional property of U. 00
The difference between any two symbols for A lies in H . In particular, then, the function p+q r+s
Ul- UO 00
Since rand s
is in H in Hi.
But also
-
q s
det U s(r+s)
are in H2,
it follows that
det U
det U has unit modulus almost everywhere on
due to the J-unitarity of U,
so det U
is an inner function.
det U were not constant, there would be a nonzero function h H2
e
(det U)H 2 .
Assuming that to be the case, let x
projection onto H of the function lation shows that x every n f; 0
h(9 (s-lqh).
is J-orthogonal to
n
and thus is J-orthogonal to N.
must be in G(A*) ,
so its first component,
aD If in
be the
A simple calcu-
and to
S Xl
is
n
S x 2 for
But that means x h,
must lie in (H2)1.
This is a contradiction, because h was presumably a nonzero vector in H2.
We can conclude that det U
is constant.
By
suitably choosing the vectors Xl and x 2 ' we may assume det U = 1. l implies p = sand q = r, so Then the equality U[*1 = U-
U=(!
~).
306
D. SARASON 00
Suppose now that <j> is any symbol for A in ball L . The U-1<j> is then in ball L00 function 1jJ We shall show that 1jJ is actually in ball Hoo , thereby completing the characterization of the class of symbols for A whose norms do not exceed 1.
(The
reformulation in terms of graphs of the result we are about to establish will be left unexpressed.
At this point graphs have
served their purpose and no longer need to be referred to.) As
rls
00
(= UO)
is a symbol for A,
The expression for 1jJ
is in H.
in terms of <j>
sep - r -r<j> +
the difference
aD,
nator,
-r<j> +
s,
s(ep-rls)
s
we have
is equal to
-r<j> + s
s
= (IrI 2 +l)/s,
rI s) + II s,
-r (<j> -
The denominator can be rewritten as 1s 12 - rs<j>
rls
is
The numerator in the preceding fraction is in H2. Isl2 = Irl2+l on
<j> -
~
(I
is thus the product of two H2
S
Because so the denomi-
which is also in H2.
12 - rs<j».
The function
functions, so it is in HI.
As it obviously has a positive real part, it is an outer function
(12, p .112) .
s,
Thus -r<j> +
being the product of two outer functions, is an outer function. The Loc function 1jJ is thus the quotient of two H2 It follows that 1jJ
functions, with the denominator being outer. is in Hoc, as desired.
To conclude this lecture we indicate how to derive an explicit expression for one choice of the vectors xl and hence for one version of the matrix function U, the operator A.
and x 2 ' in terms of
The reasoning is like that used for the analogous
purpose in the last lecture. The subspace N,
as one easily sees, is the orthogonal sum
of the subspaces H2 Ell {O}
and G(A).
lie in the smaller subspace by S*
into G(A*),
vector x
for some g
in
(c: Ell {O}) + G (A),
and hence into (H2)1 Ell H2.
in L has the form
and some scalar a..
A vector in L must actually
That
(H2)1.
x
=
(Af + a.) Ee f
S*x li",5 in G(A*)
because it is mapped Therefore, any for some f means S*x
in H2 g
G:)
To avoid some notational confusion, at
A*g
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
this point we let S1
denote the bilateral shift on L 2 and S2
the unilateral shift on H2 S*x = g ~ A*g
307
(so
S = S1
@
S2).
From the equality
one deduces, after a little fiddling, the relation S* (1 - A*A)f
aA*S* 1
2
1
which implies that (1- A*A)f
for some scalar 13. f
as 2 A*S*1 1 +
Consequently, the equality 0.(1- A*A) -1 S A*S* 1 + 13(1- A*A) -1 1 2
1
is a necessary condition for the vector to L.
Q
JJ
x = (Af + a)
@
f
to belong
The reasoning is reversible, so the equality is a suffic-
ient as well as a necessary condition.
In particular, if one sets
0.= 0 one obtains for x the vector
which is obviously strictly negative and so, after the scalar 13 has been properly adjusted, becomes a possible choice for x 2 . Taking 13 to be positive for simplicity, one obtains in this way the following formulas for the entries of one version of U: s
p
q
We note that, because
r
is in (H2)1,
the function
r
vanishes
at O. 5.
EXTENSIONS OF HANKEL OPERATORS (CONTINUED) We continue to examine the extension problem for Hankel
operators, now taking up the case IIAII = 1.
That case will be
approached as the limit of the case IIAII < 1, the method being an adaptation due to A.Stray (21) of the method of Nevanlinna used in Lecture 2 to treat the infinite Nevanlinna-Pick problem.
308
D. SARASON
We assume given a Hankel operator A satisfying IIAII = 1, and for
0 < £ < 1 we let
A£ = (1 - £)A.
From the analysis in the last
lecture we know that the set of symbols for A£ {U£1/!: 1/! E ball HOO } ,
where U£
J-unitary almost everywhere on
in ball L oo is
is a 2 x 2 matrix function which is
aD
and of the form
(!: ::),
with
r£ (0) = 0 and
We note that
00
is in ball H
I f 1/!
then the function
s£(r£1/!+ s£) lies in Hoo •
As 1/! varies over ball Hoo , the set of values taken
by the preceding function at a given point disk whose radius,
Hoo
2
-lr£(z)1
P (z), £
2 )-1.
z
of D
is a closed
is easily computed to be
Because r£(O)=O and r£fs£
is in
we find, using Schwarz's lemma, that
Standard reasoning shows that the set of symbols for A in 00
ball L
is non-empty.
In fact, one can obtain such a symbol by 00
taking a cluster point, relative to the weak-star topology of L , Moreover, a normal families argument shows that the set of symbols for A in ball L 00 is a singleton (The preceding lim p (z) = 0 for every z in D. +0 £ limit always exists because (1_£)-1 P£(z) decreases with £.)
i f and only i f
£
lim P£(z) > 0 for some z in D, we say the extension £+0 problem for A is indeterminate. By virtue of the inequality If
above, that happens precisely when
supls£(z)I £
SOll'e
z
in D.
is finite for
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
309
By Harnack's inequality (applied to the functions log s£ ), the functions
s£
remain bounded for some point of D if and only
if they are uniformly bounded on compacta in D.
The boundedness
of these functions at 0 is thus necessary and sufficient for the indeterminateness of the extension problem for A. that
II (1- A~A£)-~ 1112 ' lim 11(1-A~A£)-~1112 <
s£ (0) =
condition
We have noted
from which one infers that the 00
is necessary and sufficient
£+0
for the indeterminateness of the extension problem for A.
That
condition is easily seen to hold if and only if the function 1 lies in the range of the operator
(1 -
A*A)~.
We assume now that the extension problem for A is indeterminate, so that the functions
s£'
and hence also the functions
are uniformly bounded on compacta in D.
r£,
We can thus select a
(£0)7 tending to 0 such that both sequences (s£ ) and n ) converge uniformly on compacta in D, say to the functions
sequence
(r £n sand r, respectively.
For simplicity we write sn in place of
s£ and rn in place of r £ (and also Un in place of U£ ). n n o o n Because 1/ sn and rn/ sn belong to ball H for each n, the same In particular, the functions rand s
is true of l/s and r/s.
belong to the Nevanlinna class and so have well-defined boundary values almost everywhere on
aD.
Passing to a further subsequence, if necessary, we can assume _
that the sequence
00
(rn/s n )1
converges in the weak-star topology
of L00, say to the function cf> 0' which is clearly a symbol for A. If 1/J is any function in ball Hoo , then the functions
00
are in H , are bounded in modulus by 2, and converge pointwise in D to 1/J/s(r1/J+s).
They therefore converge to the same function relative to the weak-star topology of Loo , from which one infers that the function cf>o
+ s(r1/J+s)
310
D. SARASON 00
is a symbol for A.
This function is in ball L , being the weak-
star limit of the functions unlJi
in ball L00.
Let U be the matrix function
(! :)
seems reasonable to expect that, for lJi >
o
(and in particular >0
+
lJi s (rlJi + s)
= rls).
(defined on aD).
It
in ball Hoo , we have
UlJi
The expectation is accurate, but
the conclusion is not immediate because we do not at present have a strong hold on how the matrices Un
converge to U.
The proof
of the equality above will be indirect, the crux of the matter being the proof, given in the next paragraph, that, for any complex number
A of unit modulus, the function >
A
>
0
+ ----,-.:.,:.A_-:s(rA+s)
is of unit modulus almost everywhere.
(This will tell us, in
particular, that the matrix U is J-unitary almost everywhere.) Let function
A be a complex number of unit modulus. (1 - En)>A
so we have
00
with lJin
in ba11 Hoo •
The preceding
equality can be rewritten (1- En) A 1 ( - En) > 0 + s ( r A + s)
from which it follows that
s(d
+ s)
in the weak-star topology of L"" and hence pointwise in D. Therefore
lJi n -> A pointwise in D.
the
(1 - En) A and is in ba11 L ,
is a symbol for
(1- En)>A = UnlJin
For each n
On aD we have
(1 - En) sn>" - rn - (1 - En) rn
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
311
One can now follow word for word the proof of Nevanlinna given at the end of Lecture 2 to show that 1jJn (0)
could not converge to A
were CPA not unimodular. Let A,
as above, be a number of unit modulus.
We know that
UnA ~ CPA
in the weak-star topology of Leo and hence also in the weak topology of L 2 . Each function UnA is unimodular, and so is CPA' and
as we just proved.
"CPA" 2
1.
=
Therefore
"Un A"2 = 1
UnA ~ CPA
It follows that
for all n,
in L 2-norm •
We may think of UnA as an L 2-valued function of A, similarly for CPA' deduces that
2;i
From what has just been established one easily converges in L2 -norm to 2;i
faD unA dA
in other words, that 'in / sn converges in L 2-norm to cP o' A-I(UnA-UnO)dA 21.1 TIl- aD
conclusion implies weakly in L 2
converges in L2-norm to
21. 1 TIl-
1/ s~ converges in L 2 -norm to 1/ s 2 .
in other words,
and
Ill/snl12 ~ lIl/sI12,
1/ sn
(because
---';l>
1/ s
aD
faD CPA dA, Similarly,
).-I(cp,_cp )dA, 0
1\
The last l/sn ~ l/s
and since also
weak-star in L00), we see that
l/sn ~ l/s in L2-norm. In virtue of the preceding observations, we can assume, rn/sn ~ CPo
passing to a subsequence if necessary, that everywhere and Unl
~
l/sn
~
l/s
almost everywhere.
2
CPI
in L -norm, we can assume
Unl
l/sn(rn+s n ) ~l/s(r+s)
which implies
now follows that rn -+ r
~
CPI
Also, because almost everywhere,
almost everywhere.
almost everywhere, that
that Un -+ U almost everywhere.
almost
CPo
=
r/s,
It
and
In particular, U is J-unitary
almost everywhere. It is now a simple matter to show that the set of symbols for A in ball Loo is
{U1jJ: 1jJ E ball Reo}. That U1jJ is a symbol for eo 1jJ is in ball R has been established (for we now have
A whenever the equality
CPo + s(r*+ s)
the preceding paragraph).
= U1jJ,
For the other direction, suppose cP
any symbol for A in ball Leo. is a symbol for (1- En)CP
=
Un 1jJn
(1 - En)A
with
by virtue of the results in
For each n,
and it belongs to ball Leo, so
1/In in ball
Reo.
is
the function (1- En)CP
Passing to a subsequence
312
D. SARASON
if necessary, we can assume the sequence uniformly on compacta in D,
As n
-+
00
,
00
(1j!n) 1
say to the function 1j!.
weak-star topology of
is +
For each n,
the left side here converges in L2 -norm to
and the right side converges pointwise in D,
converges
1jJ
L oo ,
= U1jJ,
r / s,
and hence in the
1jJ
to Therefore, s(r1j!+ s) as desired. This completes our
s (r1j! + s) discussion of the extension problem for Hankel operators. The results presented in this lecture and the preceding one were originally established by V. M. Adamjan, D. Z. Arov and M. G. Krein (1),(2).
They used an operator-theoretic method, based on
scattering theory, different from the one employed here.
Their
method is, in a way, the analogue for the Nevan1inna-Pick problem of the operator-theoretic approach to the classical moment problem (3).
While perhaps less direct than the Ball-Helton approach,
it provides insights of its own and has the advantage of enabling one to treat simultaneously the cases IIAII < 1 and IIAII = 1 • The problem here Hankel operators imating L
oo
referred to as the extension problem for
is clearly equivalent to the problem of approx-
functions in the Loo -norm by Hoo functions.
An
approach to that problem which bypasses operator theory has been developed by J. B. Garnett (9).
Another is due to A.Stray in the
paper mentioned at the beginning of this lecture; it appears in modified form above. REFERENCES (1)
(2)
Adamjan,V.M., Arov,D.Z. and Krein,M.G., Infinite Hankel matrices and generalized problems of Caratheodory-Fejer and F.Riesz. Funkciona1. Anal. i Pri1ozen. 2 (1968), vyp. 1, pp.1-19. Adamjan,V.M., Arov,D.Z. and Krein,M.G., Infinite Hankel matrices and generalized problems of CaratheodorY-Fejer and I.Schur. Funkcional. Anal. i Prilo~en 2 (1968), vyp. 4, pp. 1-17.
THE NEVANLINNA-PICK INTERPOLATION PROBLEM
313
(3)
Akhiezer,N.I., "The Classical Moment Problem," Hafner, New York, 1965.
(4)
Ando,T., "Linear Operators on Krein Spaces," Hokkaido University, Sapporo, Japan, 1979.
(5)
Ball,J.A. and Helton,J.W., A Beurling-Lax theorem for the Lie group U(m,n) which contains most classical interpolation theory. J. Operator Theory 9 (1983), pp.l07-l42.
(6)
Bognar,J., "Indefinite Inner Product Spaces," Springer-Verlag, New York, 1974.
(7)
Caratheodory,C., Uber den Variabilitatsbereich der Koeffizienten von Potenzreihen, die gegebene Werte nicht annehmen. Math. Ann. 64 (1907), pp.95-ll5. Duren,P.L., "Theory of HP Spaces," Academic Press, New York and London, 1970.
(8) (9)
Garnett.J.B., "Bounded Analytic Functions," Academic Press, New York and London, 1981.
(10)
Halmos,P.R., Shifts on Hilbert spaces. J. Reine Angew. Math. 208 (1961), pp.l02-ll2.
(11)
Halmos,P.R., "A Hilbert Space Problem Book," Springer-Verlag, New York, 1982.
(12)
Koosis,P., "Introduction to Hp Spaces," Cambridge University Press, Cambridge, England, 1980.
(13)
Krein,M.G., Introduction to the geometry of indefinite J-spaces and to the theory of operators in those spaces. Amer. Math. Soc. Transl •• Ser. 2, Vol. 93 (1970), pp.l03-l76.
(14)
Nehari,Z., On bounded bilinear forms. (1957), pp.153-l62.
(15)
Nevanlinna.R., tiber Beschrankte Funktionen die in gegebene Punkten vorgeschriebene Werte annehmen. Ann. Acad. Sci. Fenn. Ser. A 13 (1919), no. 1.
(16)
Nevanlinna,R., Uber Beschrankte analytische Funktionen. Ann. Acad. Sci. Fenn. Ser. A 32 (1929), no. 7.
(17)
Pick,G., tiber die Beschrankungen analytischer Funktionen, welche durch vorgegebene Funktionswerte bewirkt werden. Math. Ann. 77 (1916). pp.7-23.
(18)
Rosenblum,M. and Rovnyak,J., An operator-theoretic approach to theorems of Pick-Nevanlinna and Loewner types. Integral Equations and Operator Theory 3 (1980). pp.408-436.
(19)
Sarason,D., Generalized interpolation in Hoo • Math. Soc. 127 (1967), pp.179-203.
(20)
Schur,I., tiber Potenzreihen, die im Innern des Einheitskreises beschrankt sind. I,ll. J. Reine Angew. Math. 147 (1917),
Ann. of Math. 65
Trans. Amer.
D. SARASON
314
pp.205-232, 148 (1918), pp.122-l45. (21)
Stray,A., On a formula of V.M.Adamjan, D.Z.Arov and H.G.Krein. Proc. Amer. Math. Soc. 83 (1981), pp.337-340.
(22)
Sz.-Nagy,B. and Foias,C., "Harmonic Analysis of Operators on Hilbert Space," North-Holland, Amsterdam, 1970.
(23)
Sz.-Nagy,B. and Koranyi,A., Relations d'un probleme de Nevanlinna et Pick avec la theorie des operateurs de l'espace Hilbertien. Acta Hath. Acad. Sci. Hungar. 7 (1956), pp.295-302.
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
Allen L. Shields The University of Michigan Department of Mathematics Ann Arbor, MI 48109
INTRODUCTION. In these three lectures we consider Banach spaces of analytic functions on plane domains. multiplication by
If the space admits the operator of
z, then it is of interest to describe the cyclic
vectors for this operator, that is, those functions in the space with the property that the polynomial mUltiples of the function are dense.
A necessary condition is that the function have no zeros;
in general it is difficult to give necessary and sufficient conditions. In the first chapter this problem is discussed in full 0enera1ity, various elementary propositions are established, and several questions are posed.
Most of this material is taken from [9].
In
the second and third chapters we consider two special Hilbert spaces 2 The space H plays a
of analytic functions in the unit disc.
special role in the study of rotation invariant Hilbert spaces. In the second chapter we study the Bergman space, as an example of the 2 situation when the space contains H as a dense subset. In the third chapter we study the Dirichlet space, which is contained in
H2
as a dense subset. 315
A. L. SHIELDS
316
vie assume familiarity wi th some basic facts about Hardy spaces (for example, inner and outer functions, and Beurling's theorem). These may be found, for example, in [14], ~
NOTATIONS. in
~
[17],
[20], or [25]
will denote the open unit disCi if
f
is analytic
then we denote the power series coefficients by
f(n):
n
~
f(z) =Ef(n)z. CHAPTER ONE:
BANACH SPACES OF ANALYTIC FUNCTIONS.
Let
G
be a bounded region in the complex plane.
say that
E
is a Banach space of analytic functions on
We shall
G
if the
following conditions are satisfied. 1.
E
functions in
is a vector subspace of the space of all holomorphic
G.
2.
E
3.
The linear functionals of evaluation at a point are
has a norm with respect to which it is complete.
continuous with respect to the norm of 4.
E
5.
If
6.
To each point
E, for each point in
G.
contains the polynomials as a dense subset. feE,
sponds a function
then
f
zf e E.
w e aG in
E
(the boundary of
G) there corre-
that has a singularity at
w.
We never use this last axiom explicitly, but without it some of the questions that we pose would be trivial. Before discussing some consequences of these axioms we turn D (_00 < ex < 00) we denote the space ex holomorphic in ~, for which
to a class of examples. of functions
f,
By
00
II fl12 For
ex < 0
(n + l)ex I f(n) 12 < 00 . 0 this norm is equivalent to
I
J~ If(re i6 ) 12(1 - r 2 )-1-ex rdrd6 (see, for example,
[53], Lemma 2).
(1)
(2)
It is easy to see that
if and only if f' e Dex _ 2 • Also, it is not difficult to ex satisfy the six axioms above. (Axiom 3 show that the spaces D ex is proved by applying the Cauchy inequality to the power series feD
317
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
for
fez)
see Lemma 1 of [53].)
i
Spaces closely related to the
spaces (0 < a ~ 1) have been considered by Carleson in his a dissertation [10, §9], and by Salem and Zygmund [39].
D
~V'e
space B.
call attention to the three values of D_l
a:
-1, 0, 1. The
is called the Bergman space and will be denoted by
The space
DO
is the Hardy space
2
H.
The space
called the Dirichlet space and will be denoted by 2 = fflf(re i6 ) 12 rdrd6 ,
{ llfll~
=L I~(~) 1 A
Ilfll~
L(n
+
Dl
D.
is
We have
~
1) If(n) 12 =
Later we shall be
Ilfll~2
+
(3)
Ilflll~·
concerned with the space
D.
This is the
space of all those analytic functions that map the unit disk onto a Riemann surface of finite area, the area being equal to Also,
feD
if and only if
fl e B.
We return now to our general situation.
As before,
denote a Banach space of analytic functions on a region that a family of analytic functions in
G
E
will
G. Recall
is said to be a normal
family if each sequence in the family contains a subsequence that converges uniformly on each compact subset of
G.
The necessary
and sufficient condition for this is that the functions in the family be uniformly bounded on each compact subset of PROPOSITION 1. PROOF.
The unit ball in
E
G.
Let
e
G.
(f,A) = fez) z By Axiom 3 these functionals are bounded.
sup{1 (f,A ) I: z € K} < 00 for each f. z ur.iform boundedness there is a constant for all
z
e
z e K.
be compact, and let
KeG
denote the functional of evaluation at z
is a normal family.
We must show that the unit ball is uniformly bounded
on each compact subset of
feE,
G.
If(z) I ~ c K
Thus
Z:
If
fn
~
f
Also,
Hence by the principle of cK
for each
weakly in
uniformly on each compact subset of
G.
z
for
II Az II ~ c K f e Ball (E) , and all
such that
K.
COROLLARY.
A
E,
then
f
n
(z)
~
fez)
A. L. SHIELDS
318
We omit the proof.
When
E
is a reflexive Banach space we
have a converse to this corollary. general case when
E
First we consider the more
is a conjugate Banach space.
this in the sense of topological isomorphism: one Banach space follows
X such that
E
We shall take
there is at least
is isomorphic to
In what
X*.
X will be fixed so that we have a particular weak* top-
ology on
E.
PROPOSITION 2.
Let
E
be isomorphic to a conjugate Banach
space and assume that for each point evaluation at
z
e
z
is weak* continuous.
G
Let
the functional of {f}C: E n
be given.
Then the following three statements are equivalent.
and
f
b)
(i)
n
f (z) -+ 0 uniformly on each compact subset of n Ilfnll ~ const.
(ii)
c)
(weak*) .
-+ 0
a)
(i)
fn(z)
For the
If
then
n
e
f
n
E
"f
n
(ii)
!lfnll
<
const.
is a reflexive Banach space and if weakly if and only if both of the fol-
-+ 0
lowing conditions are satisfied (ii)
and
G),
proof see [9, Proposition 21.
COROLLARY. {f } C E,
(z
-+ 0
G,
II
(i)
f
n
(z) -+ 0
(z
e
G) ,
const.
<
We turn now to a special class of linear transformations on E. A complex-valued function
DEFINITION. mul tiplier on M~
E
~E
if
in
~
G
is called a
C E.
we denote the operator of multiplication by
~f
(f
An
e
E).
The set of all multipliers will be denoted
application of the closed graph theorem shows that
is a bounded linear transformation on finite norm is analytic in
Since G.
1
e
E
Hence it has a
E.
we have
~
e
E
~
and so
The following result [15, Lemma III shows that
multipliers are bounded functions. PROPOSITION 3.
If
~
e M(E)
then
I ~ (z) I <
"M~
II ,
z
e G.
319
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
PROOF.
z € G,
If
evaluation at
z:
(f,A
let z
)
A denote the linear functional of z = f(z), f e E. (We write (f,A) to
denote the value of the functional f e Ball E
f. )
A at the vector
If
we have:
14>(z) (f,A ) 1 = 14>(z)f(z) 1 z
The result follows by taking the supremum on NOTATION. tions in
G;
Hoo(G) when
COROLLARY.
f.
denotes the space of bounded analytic func-
G
we write simply
~
M(E) C E
00
H.
n Hoo (G) .
Before proceeding we describe the multipliers on the spaces.
D
generally, a complete description of the multipliers from DS
a
These results are contained in [51] and [53] where, more
is given, for all
a,S.
D
a
As usual by the disc algebra
A
to we
mean the space of functions continuous on the closed unit disc and analytic in 1.
For
~
with the supremum norm.
~,
a > 1,
D
a
is an algebra (see Theorem 3 of [26],
or Example 1, §9, p. 99 of [47]), and thus
M(D )
a
= D. a
DCA. (in fact, the power series converge absolutely). a
Also, It can
be shown that the maximal ideal space is the closed unit disc (see [47] , Cor. 1 to Prop. 31, p. 94). 00
= H (For a < 0 this follows from a (2) , and for a = 0 it follows from the corresponding integral representation of the H2 norm. )
2.
For
a < 0,
3.
The case
M(D)
0 < a < 1
is considerably more difficult (see
Theorem 1.1 (c) and 2.3 of [51]). The result involves a comparison of the measure 14>' 12 (1_r)1-a rdrd6 (where 4> e M(D» with cera
tain Bessel capacities; when
a = 1
the capacity can be taken to
be the ordinary logarithmic capacity. 4.
For
a > S
we have
5.
For
a > 1
we have
C M(D S) D CM(D) CD
M(D a )
a
(see [53], p. 233). Hoo (this follows
n
from 1. and 4. above, and from Proposition 3; it can also be
A. L. SHIELDS
320
proved directly without using (that is,
6.
¢ If
e
D2 )
¢
then
4).
~
is analytic on the closed unit disc D
e
H2
e
M(D)
if and only if
result now follows since
e
¢f'
then it
spaces (this follows from 1. and 4.) .
a
¢ ¢ e M(D)
¢'
¢ e M(D) •
is a mUltiplier on all the
Indeed,
In particular, if
¢'D CB.
if and only if
(¢f)'
e
B,
for all
feD. The
B.
t\Te now return to the general theory; as before, let Banach space of analytic functions on a region Axioms 1-6 are satisfied.
A function
vector (for the operator
M
polynomial multiples
f
are dense in
the usual terminology in operator theory:
be a
that is,
is called a cyclic
acting on the space
z
of
feE
G,
E
E.
if the
E)
This agrees with
a vector
x
is cyclic
for an operator
T if the finite linear combinations of the 2 x, Tx, T x, ... are dense. Since we shall consider
vectors
several different spaces of analytic functions we shall sometimes say
"f
is cyclic for the space
function
1
e
If
f
is cyclic, then
E, by Axiom 4.
fez)
~
0
for all
G.
PROOF. vanish at
Let
e
zo
G.
The set of all functions in
is a proper closed subspace of
that is mapped into itself by QUESTION 1.
~
0
E
for all
E
that
(Axioms 3 and 4)
M . z
Does there exist a Banach space
functions for which a function fez)
Note that the constant
is a cyclic vector for every space
PROPOSITION 4.
z
E".
f
E
of analytic
is cyclic if and only if
z e G?
He note that if
E
were such a space, then the set of cyclic
vectors would form a non-empty, relatively closed subset of EVO}. f
-+ 9
{f }C E n in norm (or weakly) , and that Indeed, assume that
n to Proposition 1,
Also, the functions
f
n
(z) -+ g(z) f
n
are cyclic vectors and that 9 ~ O.
Then by the corollary
uniformly on compact subsets of G.
have no zeros in
G,
by Proposition 4.
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
Hence by Hurwitz' theorem, either g
o.
=
that
g
321
has no zeros in
G,
or
Since the latter possibility has been excluded, we see g
must be cyclic.
The only known examples of this phenomenon in general operator theory (that is, of an operator on a Banach space
X
such that the
set of cyclic vectors is a non-empty, relatively closed subset of X\{O}) are the recent examples of operators for which every nonzero vector is cylic (that is, no non-trivial closed subspace is mapped into intself by the operator) due to Enflo [16], Beauzarny [2], and Read [36],
[37],
is a Banach space and
on
X,
T
If
X
T
each
n
let
Qn
the sets
denote the set of vectors p
(depending on
is an open subset of
Qn
set.
If
X
is
be a countable base for the open sets.
n
exists a polynomial Clearly
If
is not separable then there are no cyclic
{U}
Qn
~l).
Go
is a
vectors and so the proposition is trivially true. separable, let
=
is a bounded linear transformation
then the set of cyclic vectors for
PROOF.
X
(H.S. Shapiro, see [47, §ll, p. 110]).
PROPOSITION 5.
x
(in this last reference,
X,
f)
f
For
for which there
such that
p(T)f
e Q . n
and the intersection of all
is just the set of cyclic vectors.
This completes
the proof. Finally we remark that it is easy to give examples where the set of cyclic vectors is a non-empty open set. Returning now to Banach spaces of analytic functions, it is natural to impose additional conditions on ishing in
G.
In
2
H,
is cyclic if and only if
f
besides non-van-
for example, Beurling [6] showed that f
has no inner factor.
The absence of
the Blaschke factor is equivalent to the non-vanishing of the additional condition is that
f
( Iz I
< 1).
f, and
have no singular inner factor.
In D (a > 1) f is cyclic if and only if a in the closed unit disc, or equivalently, If(z) I > c > 0
f
f
has no zeros
(4)
322
A. L. SHIELDS
(This follows from the fact that the maximal ideal space is the closed unit disc; one must also show that
f
only if it lies in no proper closed ideal.
is cyclic if and Thus the cyclic
vectors are precisely the invertible elements in the Banach algebra D.)
For
a
=1
a
condition (4) is still sufficient for
f
to be
cyclic (see [48]), but is no longer necessary (see Chapter Three). Also when
a
< 0
(4) is sufficient for cyclicity (this follows
from Proposition 9) . As before let functions in E
G.
E If
denote a general Banach space of analytic feE
we let
of the polynomial mUltiples of
only if
[f] f.
denote the closure in
Thus
f
is cyclic if and
[f] = E.
PROPOSITION 5.
Let
f,g e E
and let
be a polynomial.
p
Then
{Pn }
l.
p[f] C
2.
If
g e
[ f]
then
3.
If
g e
[f]
and
4.
f
[f]. [g]C [f]. g
is cyclic then
such that
6.
is cyclic i f and only if there exist polynomials
p f -+ 1 (in norm) . n f is cyclic if and only i f there exist polynomials
(weakly) • Pn f -+ 1 If E is reflexive then
f
(z e G),
and
PROOF.
(ii)
II
{p} such that n p f" < const. n -
(i)
p
n
(z)f(z) -+ 1
1. and 2. are obvious and 2. implies 3.
{p } exists with Pn f -+ 1, then n 3. ; the converse is trivial. To prove 5. note first that i f
are polynomials Conversely, i f
{p } with Pn f -+ 1 n Pn f -+ 1 weakly then
of the polynomial multiples
of
f.
closed if and only if norm closed, so
{Pn }
is cylic if and only i f
there exist polynomials
by
is cyclic.
such that 5.
if
f
1 e
f
[f]
As to 4.,
and so
f
is cyclic
is cyclic then by 4. there in norm, and hence weakly. is in the weak closure
1
But a subspace is weakly 1
e
[f]
as required.
Finally, 6. follows from 5. by the corollary to Proposition 2.
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
323
Hote that 4., 5., and 6. remain true if any other cyclic vector PROPOSITION 6.
II fl12 -<
b)
if
f
PROOF.
is replaced by
g.
If
are two Banach spaces of analytic
functions in the same region a)
1
G,
E1C E 2 ,
and i f
then
for some c > 0 and all f € E l ; c "f" 1 is cyclic for E l , then i t is cyclic for E 2 ·
By Axiom 3 the point evaluation functionals are con-
tinuous with respect to both norms. identity map of
into
El
bounded, which proves
Using this one shows that the
has a closed graph and hence is
a).
Part b) now follows from 4. of
Proposition 5. If
COROLLARY 1. f
0
< (). <
I,
and if
f € D
().
is cyclic, then
is an outer function. PROOF.
E2
2
=H
•
The converse to this corollary is not valid (see Chapter 3). Several of our results depend on an additional axiom. that
M(E)
denotes the set of multipliers on
AXIOM 7.
M(E) [f] C
[f],
all
E.
feE.
This is equivalent to requiring: QUESTION 2.
Recall
M(E)f
c: [f].
Is Axiom 7 a consequence of Axioms l-6?
This seems unlikely though at present we have no counterexample; see [9, pp. 297-298] for further discussion. PROPOSITION 7.
Each of the spaces
D (-00 < a < 00) satisfies a
Axiom 7. PROOF.
Fix
a
and let
¢ e M(D). ().
By Theorem 12(iii) of
[47] there is a sequence of polynomials (the Fejer means of the partial sums of the power series for norm, for each
f e D(). (i.e.,
operator topology).
Hence
{Pn}
¢f e
¢) such that converges to
f
-+
¢f
E
in
in the strong
[f], as required.
From now on we shall assume that our space Axiom 7.
Pnf
satisfies
324
A. L. SHIELDS
PROPOSITION 8. [¢f] C
(i) (ii)
feE
and
e
¢
M(E).
Then
If] () [¢]
I f also
(iii)
Let
f
is cyclic, then
Finally,
f
[¢f] = [¢1.
is cyclic if and only if both
f
and ¢
are cyclic. PROOF.
(i)
From Axiom 7 we have
[¢f1 C. [fl.
Now let
{p} n ¢f e [¢].
be polynomials such that
Then p ¢ -+ ¢f, and so Pn -+ f. n {p } Let be polynomials such that p f -+ l . Then n n and so ¢ e [¢f1.
(ii) Pn¢f
¢,
-+
(iii)
If
f
[¢f] = I¢l = E. [f]
=
I¢l = E.
and
are both cyclic then from (ii) :
¢
[¢f] = E
Conversely, if
then from (i) we have
This completes the proof.
In Chapter Two we will need to apply this result to a topological vector space
E
of analytic functions (see Proposition 16).
Our seven axioms are all satisfied except for axiom two: the topology is not given by a norm and is not metrizable. {p } n
we interpret
However, if
in the proof above as a net, rather than as a
sequence, then the proof remains valid (see [23, Chapter 21 for a discussion of nets).
Actually, in the application to Proposition
16 it will be sufficient to use sequences.
In the next two corollaries
G
we shall consider the functions no zeros in
G.
Of course
¢a
¢a
will be simply connected, and (a > 0) where
¢ e M(E)
has
is not uniquely determined, but
this causes no difficulty since the ratio of any two determinations of
¢a
is constant.
COROLLARY 1.
Let
be simply connected. If
¢a
¢ e M(E) Assume that
is cyclic for some
have no zeros in ¢a e M(E)
a ~ b,
then
G
for all
and let
G
a > b > O.
¢a is cyclic for all
a > b.
PROOF. above. Now let Then
Let
¢a
Similarly, c
~
b
be cyclic. ¢na
Then
¢2a
is cyclic by (iii)
is cyclic for all positive integers
be given, and choose
n
such that
n.
na - c > b.
¢na = ¢c¢na-c and both factors are cyclic by (iii) above.
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
COROLLARY 2.
Let
cp e H(G)
and let
some number
G
be simply connected, let
have no zeros in
a > 0,
325
cpa
then
G.
If
cpa
is cyclic for all
M(E)
00
=
H (G) •
is cyclic for a >
o.
We now pose two test questions about cyclic vectors in a general Banach space QUESTION 3. > Ig(z) 1
E
of analytic functions in a region G (see [49]) •
If
for all
f,g e E,
z e G,
if
g
then must
is cyclic, and if f
If(z)
1
be cyclic?
Since the constant function 1 is always a cyclic vector, we have the following special case of this question. QUESTION 3'. must
f
If
feE
and
If(z)
> c > 0
1
for all
z e G,
be cyclic?
QUESTION 4.
f, f- l e E
If
must
f
be cyclic?
This question (for the Bergman space) was posed in [47] Question 25' on p. 114).
Harold S. Shapiro [41] used the term
"weakly invertible" in place of cyclic. then be rephrased as follows: tibility?
(see
The above question could
does invertibility imply weak inver-
From Proposition 8(iii) we see that the answer is affirm-
ative if, in addition, we assume that 1 = ff- l is always cyclic .
f
is a multiplier.
Indeed,
..,
Recently F. A. Samoyan [40] gave the first example where Question 4 has a negative answer.
No examples are known where
Question 3 has a negative answer.
There is however one common
situation where Question 3 has an affirmative answer. PROPOSITION 9. >
Ig(z)
1
for all
PROOF. Hence
Since
g = CPf e
[f)
If
M(E) = HOO(G) ,
z e G,
cp
=
g/f
and if
g
if
f,g e E
is cyclic, then
> 1,
f
If(z)
by Axiom 7, and so
f
then, as remarked earlier,
f
1
is cyclic.
is bounded it is a multiplier on
E.
is cyclic.
We now consider Questions 3 and 4 for the spaces 0.
with
D. 0.
When
is cyclic if and only if it
has no zeros in the closed unit disc (recall that the functions are continuous on the closed disc).
It follows easily that both ques-
tions can be answered in the affirmative.
A. L. SHIELDS
326
0 < a < 1; see Chapter Three
Both questions are open when for partial results when
a = 1,
see also [9]
(Theorem 1 and
Corollary 1, as well as Corollary 2 to Theorem 2) . For
a < 0
Question 3.
Proposition 9 gives an affirmative answer to
For
a
Question 4 can also be answered in the affirmative: i f f and f- l are both in H2 then they are both outer functions (indeed, i f either f or f- l had a non-trivial =
0
inner factor, then so would
a < O.
f-lf = 1) .
Question 4 is open for
This leads to another interesting problem, where for sim-
plicity we specialize to the Bergman space then for
If
feB
r = Izl < 1:
2 If(z) 12= IIf(n) (n+l)-l/2 (n+l) 1/2 z n 1 Hence i f
B (a = -1).
f- l e B
then
.::.lI f ll 2
II f-lll- l
1f (z) 1 .::.
(1_r2)-2.
(1_r 2 ).
This suggests
the following question, which was first posed by H.S. Shapiro (see the Remark following Theorem 5 in [43]). QUESTION 5. c, k > 0
If
(and all
feB
and if
~),
z e
If(z) 1 > c(l-Izl)k
then must
f
for some
be cyclic?
As noted above, an affirmative answer to this question would imply an affirmative answer to Question 4 for the Bergman space (see Chapter Two for further information) • Along these lines one can pose similar questions for any Banach space of analytic functions on a bounded plane domain sider continuous functions
cp
on
is positive on
G
QUESTION 6.
Let
G.
feE
cp
on
and zero on E
(the closure of
G
G)
such that
aGo
be a Banach space of analytic functions
cp,
Does there exist a function satisfies
We con-
G.
If(z) 1 > CP(z)
for all z e G,
Such functions exist for the by Proposition lOb. below, i f
D
a
f e D
a
fies a Lipschitz condition of order has a zero on the boundary, say
as above, such that if then
f
spaces when
a > 1.
(1 < a < 3)
then
b = (a - 1)/2.
fell = 0, then
If(z)1 = If(z) - f(l)I.::.clz - lib,
Izl .::. 1.
is cyclic? Indeed, f
sat is-
Hence, if
f
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
¢(z)
Hence if satisfies
(1 _ Izl)a
=
I fez) I .:: ¢ (z)
o
where 6,
in
D
a
then it also works for
a < b,
<
then
closed unit disk and thus is cyclic.
f
if
H2
Of course if
for all
8 > a , since
Iz I
do not exist for the
(5)
< 1), ~,
and
c
is a posi-
Also, one can construct singular measures
w(8)/8
tends to infinity arbitrarily slowly.
and [44, p. 265].)
Thus, given any function
there exists a singular inner function for
¢
is the modulus of continuity of
tive constant.
§5]
works for
~,then
w
which
¢
is a singular inner function with associated singular
(r =
where
f e Da
Indeed, H.S. Shapiro has shown (see [41], p. 164) that
S(z)
measure
and if
has no zeros on the
In contrast to this, such functions space
327
S
¢
such that
~
for
(See [l8; as above, IS(z) I .:: ¢(z)
I z I < 1. Question 6 is open for the spaces
intervals:
~
0 < a
1,
and
PROPOSITION lO.a. I g (z) I where
c b.
~ c
If
a
c
then
8,
not on
I I < 1),
(z
g.
(1 < a < 3), then
If(z)-f(w)1 < cllfll where
a < O.
Izl) -(1-8)/2
II g II 8 (1 -
feD
in the remaining
a
g e D8 (8 < 1)
depends only on If
D
a
depends only on
Iz_ w l(a-l)/2 (Izl < 1, Iwl < 1), a,
not on
f.
For the proof see 19, p. 278]. CHAPTER TWO:
THE BERGMAN SPACE.
As noted in Chapter 1, B,
the Bergman space
M(B),
the space of multipliers on
is just the space
analytic functions in the disc.
00
H
of all bounded
Then from Proposition 8(iii) we
00
see that if if both
¢
¢ e H and
f
and
feB,
are cyclic.
then
¢f
is cylic if and only
Hence it is of interest to learn
328
A. L. SHIELDS 00
which
H
functions are cyclic in
Proposition 6.
Hoo
However
B.
We would like to apply
is not separable and hence has no
cyclic vectors (in its norm topology). Shapiro [43, p. 325], now follows from
If
e
f
inner function, then 5
is contained in
B.
The next result
Proposition 6 and 8(iii).
COROLLARY.
and only if
Hl
However as noted by H.S.
f
Hl
is outer and if
is cyclic in
5
B, and
is a singular fS
is cyclic if
is cyclic.
Before discussing singular inner functions we introduce a class of "thin" subsets of the unit circle that playa basic role in the theory. By a BCH set (sometimes called a Carleson set) we mean a compact subset
K
of
such that
a~
K
has Lebesgue measure
0,
I
II I (-log II I) < 00. Here {I} are the disjoint open n n n arcs in the complement of K, and I· I denotes normalized and
Lebesgue measure.
These sets were introduced by Beurling [5] in
1940, studied by Carleson [11] in 1952, and rediscovered in a completely different context by Hayman [19] in 1953. NOTATION. in
00
A
will denote the class of analytic functions
all of whose derivatives are continuous on the closed disc.
~
PROPOSITION 11. a) an outer function tives vanish on
f K;
in
KC
If 00
A
a~
is a BCH set then there is
such that
in addition,
f
f
and all its deriva-
has no other zeros in the
closed disc. b)
If
f
satisfies a Lipschitz condition of some positive
order in the closed disc (and is analytic in the open disc), then the boundary zero set of
f
is a BCH set.
Part b) was observed by Beurling
[5] in 1940.
Part a) was
proved, for functions with a prescribed finite number of derivatives continuous on the closed disc, by Carleson Ill] in 1952. The result was later extended to
00
A
by several authors [27],
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
[35], and [52].
329
The last of these papers also describes the
zeros inside the disc. COROLLARY.
The union of two BCH sets is a BCH set.
It can also be shown that any closed subset of a BCH set is a BCH set (this follows fairly easily from the definition of BCH set) • We turn now to singular inner NOTATION.
5
functions.
denotes the singular inner function coming
].J
from the positive singular measure THEOREM 1.
a~.
on
].J
A necessary and sufficient condition that the
singular inner function
5
be cyclic in
].J
B
is that
put no
].J
mass on any BCH set. The necessity was proved by H.S. Shapiro [43] in 1967.
The
sufficiency was obtained independently by B.I. Korenb1um [31] and J. Roberts [38] about 1979.
Roberts' proof was unpublished for
several years but was available in an unpublished exposition by Joel Shapiro [46].
Korenb1um showed that the result is an easy
consequence of the theory developed in his earlier papers [29] and [30].
\ie shall give
H.S. Shapiro's proof of the necessity
since we shall need the argument again in what follows. PROOF OF NECESSITY.
Let
g
00
be a
function on
C
E (Inl
generally, we only need to require that
_00
+ 1)
a~
(more
Ig(n)1
2 <00)
and define A (f) g
feB.
(6)
o
This defines a continuous linear functional on if
f
e
H1
Ag(f)
B (see (3». Also,
then 1
21T
Jor
21T
fee
is
)g(e
is
This may be seen by replacing and using the fact that
fr
~ f
)dS. f
(7)
by in
f,
L
1r
where
norm as
f
r
f(rz) ,
(z)
r t 1.
Now suppose we are given a positive singular measure puts mass on some BCH set
K.
We must show that
[5
].J
]
"I
].J
B
that (as in
330
A. L. SHIELDS
Lecture I, f).
[f] denotes the closure of the polynomial mUltiples of ~
We first assume that
K.
is supported on
00
By Proposition 11 there is an outer function ¢ in A (not ¢ (j) identically zero) such that vanishes on K, for all j. Let p(z) = dist(z,K). constants
Then by Taylor's theorem there are positive
c = c(j,k)
such that
( [wi For
m
0, I, 2, ... ).
j ,k
1;
1,2, ... , let
={wmj(W)
gm (w)
(w
5 (w), )l
e
i8
¢ K) (8)
0,
(w
e
K) .
Calculations based on Leibniz' formula for higher derivatives of a 00
e c for each m, and thus m defines a continuous linear functional on B. Let g
product show that
shall show that that
[5]
from (7)
~
B,
Al = 0
on
(with
Let
¢
is analytic.
Assume that Thus
h = 5¢
¢ = 5h,
and
A
=
gm We
5 = 5. )l
g = g ) 1
n n --n+l Al(z 5) =Jw 5(w)g(w)dm(w) =Jw ¢dm = 0 since
m
and that Al ~ 0, thus showing i8 ,dm(w) = d8/2rr. Then w = e
[5],
as required.
A
Thus Al =0
o.
Al 00
is in ¢
H
Then
(n > 0)
on
[5]. n n+l0 = Al (z ) =Jw ¢5dm,
n > O.
But this is a contradiction since
is an outer function whereas
5
is a non-constant
inner function. Finally, if where
is not supported on
~
is supported on
K
and
o (K) =
= v +
K,
then
~
O.
Also,
[5 ] C
0
[5 ]
v (this is trivial, it also follows from Proposition 8i), and we [5 v ]
have seen that
~
B.
)l
This completes the proof.
Hote that the only special property of the space
B
that was
used was the fact that the Taylor coefficients do not grow too fast, so that (6) defines a continuous linear functional on whenever
g
~
00
c.
B
Thus the result is valid for many other spaces
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
besides
B.
00 A.
space to
It is natural in this context to introduce the dual
-00 A
NOTATION. in
~
331
denotes the space of analytic functions
for which there exists
c = c(f) > 0
and
f
k = k(f) > 0
such that
(z
e
~).
It can be shown that this is equivalent to the following: there exists
~
If(n) I
b
= b(f)
b(n+1)j,
> 0
and
j
n > O. Hence if
=
j(f) > 0
g
e
COO
such that
and
e
f
A- oo
then the
series defining A (f) in (6) is absolutely convergent. (If we g 00 topologize A in the obvious way with a countable family of _00 seminorms, then A may be identified with the dual space. Since the space of distributions is the space of continuous linear 00 -00 functiona1s on C , we may regard A as the space of "analytic distributions" • ) -00 A
If we give
the weak topology from this pairing, then
Shapiro's proof above really establishes the following stronger result. PROPOSITION 12. -00 cyclic in A then
If the singular inner function )l
S
)l
is
puts no mass on any BCH set.
The converse is also valid; this follows from Theorem 1. NOTATION. in
~
AP a.
denotes the space of analytic functions
f
for which (O
circle
M (f,r)
p Izl = r.
denote the mean of order Then for
0 < p < 00
and
p R < 1
of
f
on the
we have
(1 (1 - r) a. dr. Ilfil P = (1 M (f,r)P(1- r)a. rdr > RM (f,R)P ~ p,a. -b p p Therefore
M (f,r) = 0(1 _ r)-(l+a.)/p. p
By a theorem of Hardy and
Littlewood (see Duren [14], Theorem 5.9, p. 84) we have M (f,r) = 00 = 0(1 - r) - (2+a) /p, and so APe::: A- oo • It can be shown that a.
332
A. L. SHIELDS -00
A
=U
p>O
Also,
A2 a.
- -a.
;
that is,
00,
we omit the proof.
PROPOSITION l3. cyclic in
(p fixed).
a. >-1
i f and only i f L If(n) 12 (n+l)-1-a. <
f € A! D 1
u
A- OO
AP (a. fixed), a.
If the singular inner function
S
AP
J.I
is
for some p,a., then J.I puts no mass on any BCH a. Conversely, if this condition is satisfied then S is
set.
J.I
AP for all p,a.. a. PROOF. The first part is a consequence of what has been
cyclic in
said above; the converse was established by Korenblum [31] and Roberts [38]. We now use Theorem 1 to establish an interesting factorization theorem for inner functions (Theorem 2 and Proposition 15) that was obtained by Joel Shapiro (unpublished) on the basis of the Korenblum-Roberts Theorem (Proposition 13).
The following
two lemmas and Proposition 14 are due to P. Ahern (unpublished). All this material is taken from Joel Shapiro's unpublished exposition of the Korenblum-Roberts result [46].
In what follows J.I
denotes a finite, positive, Borel measure on the unit circle, singular with respect to Lebesgue measure.
The results will be
stated for the Bergman space, but they are valid (with the same proofs) in all the spaces
AP • a.
They are also valid in
A-00 (with
the weak topology. LEMMA 1. f €
[S ] J.I
n HI,
and define
is supported on a BCH set
J.I
then
S
J.I
divides
K,
and if
f.
We follow the proof of the necessity of Theorem 1,
PROOF.
gm
will denote the corresponding linear A m Just as in that proof we have A = 0 on [S ] , m J.I
by (8) ; B.
functional on
Therefore
m > l.
Thus
If
0
A (f) =J m
h
fepS
J.I
i"gm
(on eM)
outer we see that
S
u
= J fep
-
is in
m w HI,
must divide
(m and f.
~
1) •
hS
J.I
fep.
Since
ep
is
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
LEMMA 2. Then
~
Let
Let
is concentrated on a countable
~o
~l
union of BCH sets, and PROOF.
puts no mass on any BCH set.
sup{~(K):
b
d~.
be a positive finite Borel measure on
where
~ = ~O + ~l '
333
K
a BCH set}.
Since a finite
union of BCH sets is a BCH set (see the corollary to Proposition 11) Kl c: K2 C· .. wi th
we may choose BCH sets H
UKi
=
Then
,
~O (E)
and let
is concentrated on
~o
n H)
~ (E
=
H,
~
(K i ) ->- b.
Let
for each Borel set
and
~l
=
~
d~.
puts no mass
~O
-
E C
on any BCH set. DEFINITION. q
An inner function
q
is said to be B-inner if
divides every inner function in Iq]. For example, Blaschke products are B-inner.
This concept
seems to depend on B (since, in this chapter,
Iq]
denotes the
closure in
q).
However
if
B
B
of the pOlynomial mUltiples of
is replaced by any of
example, or by
A
the spaces
A;,
for
with the weak topology, then the following
result would still be valid. PROPOSITION 14. ~
if and only if PROOF.
~.
and
~.
J J
First assume that
UK.
~
Also,
IS ] ~
is inner. 00
qj ->- qo e H as desired.
,
J
(z)
c.. [S.]. q
e
->- S
~
(z)
uniformly on compact subsets of
J
IS] ~
be inner.
'I
o.
By Lemma 1,
q
=
q.S., J J
where
By passing to a subsequence we may assume that uniformly on compact subsets of
countable union of BCH sets, and let ~l
Kj ,
J
For the converse, assume that
Thus
is concentrated on a countable
We may assume that
1.
s.
Then
Now let qj
is B-inner ~
(E) = ~(E nK j ). Then ~j is supported on the BCH set t~. Let S. denote the corresponding singular inner
function. ~.
S
is concentrated on a countable union of BCH sets.
union of BCH sets: Let
A singular inner function
Let
and
~
~.
Thus
q = qos
is not concentrated on a ~ =
~o
+
~l
as in Lemma 2.
denote the corresponding inner
A. L. SHIELDS
334
functions.
Then
sequence of polynomials PnS~ ~ SO·
Thus there is a
is cyclic by Theorem 1.
Thus
{Pn}
e
So
such that
PnSl
~
1.
Hence
is not divisible by
IS~],
S. ~
This completes the proof. THEOREM 2.
Every inner function
where
q
can be factored as is cyclic.
is B-inner and
q.
1.
The factori-
ation is unique up to a multiplicative constant of modulus one. More precisely,
v
(or constant), and sets;
and
S
= So'
qc
PROOF.
qi = bS v where b is a Blaschke product is concentrated on a countable union of Bcn
Let
q
= bS,
puts no mass on any BCH set.
where
cyclic.
Let
S = = bS O'
q.
1.
To see that =
where
b%,
e
qO·
We have
p
bqq
(since
So
is inner. (since
[SO] is
S
. c
is B-inner
So Then
c is B-inner, let
qi qo
=
q
By Lemma. 2, Proposition 14, and
where
SOSI
is a Blaschke product and
b
is a singular inner function.
Theorem 1, we have
p
a
where
qc
e
p
dnd
be inner.
[qi]
Thus we must show that [qi] C. [SO]) , and so
B-inner) •
Hence
is
Sl
is cyclic.
divides
So
divides
So So
Then
divides
qo' as
required. For the uniqueness, assume that the functions are inner, i
= 1,2.
Then
[q]
=
[II]
I.
[I 2 ].
other (since each is B-inner), S2 = CS l •
=
IlS l
are B-inner, and
1.
=
q
Thus
II
=
1 2S 2 ,
S.
are B-cyclic,
1.
and
where all
12
divide each
Ic I
and so
= 1.
Hence
This completes the proof.
D. Luecking in 1979 (unpublished) established a weaker form of the factorization theorem without use of the Korenblum-Roberts result.
Namely, he showed that if
S
is a singular inner function
then there are singular inner functions such that of
S2
Sl
then
measure for
is B-cyclic, and if S3 S2
S3
Sl'S2
with
S
is any singular divisor
is not B-cyclic; in addition, the singular is singular with respect to any measure whose
associated inner function is cyclic.
However, he was not able to
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
335
AP spaces. a It would be desirable to find an elementary proof of Theorem 2 show that this factorization was the same for all
(that is, independent of the Korenblurn-Roberts result), since later results of interest, notably Proposition 16 and its corollaries, depend on this factorization.
It would not be necessary
to describe the factors (in terms of countable unions of BCH sets, etc.) but merely to prove the existence of such a factorization. Finally we note that this result completes the theory for HP (0 < P < 1)
that was initiated in [15, §6J. qHP
an inner function then mUltiplication by closure,
[qHP ] ,
z.
q
is
is a closed subspace invariant under HP
Since
is not locally convex, the weak
may be larger than
w
First, if
qHP .
However it is still
a closed invariant subspace and so, by Beurling's Theorem (which holds also for
~
p < 1), there is an inner function where
Thus
such that
is an inner function;
is called the weak inner factor of
It was conjectured in
q.
[q HPJ = HP . Also, c w those authors asked whether the notions of weakly inner and weakly [15] that
qc
is "weakly outer", that is,
outer are independent of
p,
0 < p < 12.
The following result
establishes this conjecture and shows the independence of PROPOSITION 15.
p.
An inner function is weakly inner if and
only if it is B-inner; it is weakly outer if and only if it is B-cyclic.
Every inner function
q
where
is weakly inner and
qo
has a factorization is weakly outer.
q=qiqO
This factori-
zation is unique (up to multiplication by a constant of modulus one) , and
qi = c~, PROOF. Fix p,
[15, Theorem 7]
E
Icl = l. let
contains
(lip) - 2,
a
HP
1
and let
E = A . By a as a dense subspace and has the
same dual space (that is, a linear functional on
HP
if and only if it has a continuous linear extension to
is continuous E).
shows that p-weakly outer and E-cyclic are the same concept.
This But
we have already seen (Proposition 13) that the condition for cyclicity of an inner function is the same for
E
and
B.
336
A. L. SHIELDS
If
q
is inner let [q]E
nomial multiples of
q.
By
denote the E-closure of the poly-
[15, Lemma 8, p. 55] (9)
Now let
q
inner then
I € qHP
and so
and hence B-inner.
~
€
[q]E
q
If
I
Thus
Thus
q =
c~
q
~.
divides
where
€
[q]E
q
is
is E-inner
be B-inner and hence
be its p-weak inner factor:
by (9), and so
q.
divides
Conversely, let
~
E-inner, and let Then
qH P = [qH P ] . w q divides I.
be weakly inner:
Ic I = 1,
[qHP]w =
~HP.
But trivially, and so
q
is
weakly inner. The existence and (essential) uniqueness of the factorizatioI where
is weakly inner and
now follows from Theorem 2.
by Proposition 8ii, since
~.
qi = qo
is
Then by (9)
qi € ~HP,
This
is weakly outer,
Finally, we must show that
First note that [q]E = [qi]E E-cyclic.
qo
and
~ €
[qi]E·
Therefore
and
divide
\
each other (since
q.
1
is E-inner).
This completes the proof.
We now show how the factorization result (TheoreM 2) can be used to solve a problem raised in [1].
These results (Proposition
16 and its corollaries) are due to P. Bourdon [7J, who stated them AP spaces. Recall that a holomorphic function f in a is in the Neval~nna class N if it is equal to the quotient
for the ~
of two bounded analytic functions. b
Thus
f = bS 1 F l /S 2 F 2
where
is a Blaschke product,
S.
are singular inner functions with
no common inner division,
F.
are outer functions, and
i = 1,2.
exists
n
c,k > 0
A
with
First, a sufficient (and also necessary)
condition for a sequence (z) -+ 0
IF·I -< 1, 1
-00
To do this we state without proof some pro-
perties of this space.
f
1
We wish to apply Proposition 8 to the space
the weak topology.
(ij
1
{f} to converge to zero is that n uniformly on each compact subset of ~, (ii) there such that
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
If (z) I
c(l -
<
n
Izl)k
(Izl <1,
-co
Second, each element of
337
A
n=l, 2, ... ).
is the limit of a sequence of poly-
nomials (the Fejer means of the partial sums of the power series) . Third, i f
f
-+ f,
n
if
PROPOSITION 16. above, then
in
A
then
f e N nA
Let
is cyclic in
S2
PROOF. -co
-co
g € A
fng -+ fg.
-co
If
f = bS l F l /S 2 F 2
as
B.
For this proof let
[g]
denote the weak closure,
of the polynomial mUltiples of
weakly closed ideal generated by g).
g (this is just the
We factor
Sl
and
S2:
...,here
S,S are cyclic (that is, the u c singular measure puts no mass on any BCH set) and Sv'Sd are A
-inner.
Also, recall that 2
(since they are cyclic in
Thus
Sd
bS l .
divides
Fl
H).
and
Since
as above, then PROOF.
f
feN
S2
and
Sl
Sd = 1
nB
A
have no (non-trivial)
as was to be proved.
has no zeros and if
is cyclic i f and only i f
From (10) we have
-co
are cyclic in
Then by Proposition 8 we have
common inner factor, this forces COROLLARY l . I f
F2
[S ] = [f] 1
Sl (since
f= SlFiS2F2
is cyclic. S
d
= 1) , and
the result follows. COROLLARY 2. some
c,k> 0
and all
PROOF. ment that
f
If
f € N
n B,
Izl < 1,
The lower bound on -1
-co
eA.
if then f
If(z) I > c(l f
is cyclic in
Izl)k for B.
is equivalent to the state-
If
f = SlF l /S 2 F 2 , then Sl is cyclic, as in seen by applying Proposition 16 to l/f. The result now follows from the previous corollary. Note that this gives an affirmative answer to Question 5 in
case
f € B
is also in the Nevanlinna class.
This result was
also known to Korenblumi a special case had been proved earlier [1], where the hypothesis condition that the range of
feN f
was replaced by the stronger omits a set of values of positive
338
A. L. SHIELDS
logarithmic capacity. f,f- l e B,
if
Berman, Brown, and Cohn 14J showed that
and if
feN,
then
f
PROPOSITION 17 (Shapiro [42J). Borel measure on ~,
~
Let
f
and satisfy, for some
{p} n
PROOF. tiples of if
f
P f n
Let
[f]
in
L2(~).
, f
r}. 2
L
in
m(r)
-s
dp
<00
Then there exist polynomials
(~).
It will suffice to prove:
< min(t,s/2) ,
f(O) = 1,
fa e
and so
[f] ,
m(r)
the n-th Fejer mean of the power series of Ipn(z)I .::..max{lf(w)l- t : Thus
,
denote the closure of the polynomial mul-
0 < a < 1, 0 < t
We assume that
00
Izl
1
+
be a positive finite
s > 0,
m(r) = rnin{lf(z) I: such that
~
Let
B.
be holomorphic and non-vanishing in
<
where
is cyclic in
f
-<
-t
n
Let
l.
.
Iwl.::..lzl} .::..m(r)-t
Ip (z) - f-t(z) I.::.. 2 m(r)-s/2.
f a- t e
then
[f].
Pn denote
Then <
-s/2 m( r ) .
Therefore
!Ip fa_fa-tI2d~ =f If 2alp _f- t I 2 < 4! IfI2am(r)-sd~. n
n -
Ifl 2a .::.. 1 + Ifl2
Also,
and so the last integral is finite.
Now
by the dominated convergence theorem the integral on the left tends to zero, as desired. COROLLARY. and all
If
some
If(z) I > c(l -
Izl)k
for some
c,k
and if - r)
for
feB,
s > 0,
-s
then
dxdy < f
(ll)
00
is cyclic in
B.
By far the most important progress towards describing the AP spaces was made by Korenblum in a his two fundamental papers [29], [30]. In the first, among other cyclic vectors in
B
or the
_00
things, he gave a complete description of the zero sets of functions. A-
oo
,
A
In the second he described all the closed ideals in
and desc.clbed the cyclic vect.<.)rs.
We shall say a few words
about this work since it seems to offer the best chance for a complete description of cyclicity in B.
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
63
Let
339
denote the family of those Borel subsets of the
circle whose
closure is a BCH set.
plementary open intervals
F
is a BCH set with com-
K(F) = ElI 1(1 - 10giI I), II n where the vertical bars denote normalized Lebesgue measure. DEFINITION. a:
63 -+
{In}
If
then
A kappa singular measure is a function
such that
JR,
1.
olF
is a finite Borel measure on F, for all
2.
There exists
c > 0
with
10 (F) I 2 CK (F), for all Fe BCH. -00
Korenblum associates to each kappa singular measure
f e A
~
(f
0) a non-positive
of.
(Korenblum [30, Corollary 1.1.1]).
THEOREM 3. have no zeros in
Then
~.
F e BCH.
f
-00
is cyclic in
A
Let
-00
f e A
if and only if
of = O. CONJECTURE. no zeros in
~
feB
and
of =
is cyclic in
B
if and only if
I f(z) I > c(l
-
-
I z I) k
has
feB
and
o.
This would prove H.S. Shapiro's conjecture: if
f
for some
c,k,
Indeed the hypothesis is equivalent to:
if
then f is cyclic in B. -00 f- l e A Such an f
_00
is cyclic in
A
and hence
of = 0
by Korenblum's result.
Korenblum has informed us that he can prove the above conjecture in case if
f
is in a smaller Hilbert space, more precisely,
Illfl2(1 - r)-E <
00
for some
corollary to Proposition 17. sort of result is true when
E > O.
This generalizes the
It is not clear, however, if this f
is in a smaller
1
AP
2
state a specific question we note that AO c:: A2 lation shows that ~ 2 Ilfll~,2= Illfl2(l - r)2 < cElf(n)1 /(n If(n) Irn 2 Ml(f,r)
Also, on
rdr
we have:
by Cauchy's inequality.
If(n) I/(n + 2) 21Iflll,0.
ex
space.
To
Indeed, a calcu-
Integrating
By the Fejer-Hardy-
Littlewood inequality (sometimes called Hardy's inequality; see [14, p. 48], and for some historical remarks see [50]), we have ~
n
E If(n) Ir /(n + 1) 2 nM1(f,r).
340
A. L. SHIELDS
Integrating
rdr: l:lf(n) I/(n + 1)2 2. c Ilflll,o.
the previous result we see that gUESTION 7. must
f
If
e
f
1
II f II
Using this and
2. c II fill , 0 as required.
2,2
has no zeros in
AO
and if
/',
f
-1
-00
eA
,
2? A2 ·
be cyclic in
We conclude with a few remarks about the invariant subspaces of
B.
Those are the closed subspaces
M CB
such that
zMCM.
The collection of all such subspaces forms a lattice (partially ordered by inclusion), which we denote by
Lat(M ). z it was an open question whether two such subspaces
o.
intersect only in
At one time
(i {O})
could
It follows from Beurling's theorem that 2
this cannot happen in
However C. Horowitz showed in his dis-
H •
sertation [21] that there are two Bergman zero sets whose union is not a zero set, and thus two non-trivial invariant subspaces can indeed intersect only in
O.
More recently, Bercovici, Foias, and
Pearcy have shown that the Bergman lattice is much more pathological than anyone had suspected (see [3, Chapter 10]). positions contain
some of their results.
lattice of all closed subspaces of PROPOSITION 18.
Let
The next two proLat B
denote the
B.
¢: Lat B + Lat M z that is injective, increasing, preserves closed spans, and has the
following property:
There is a function
if
if
{E} C:Lat H, n
nE
n
=
then
{O},
n¢(E ) = {O}. n
They derive a number of corollaries including the following. COROLLARY. Lat Mz
such that
There exists Ea
PROPOSITION 19. exist that
E,F C Lat M, z zE CF. COROLLARY.
codimensions > 1
n Eb
=
x
{O}
for
a
Given an integer with
FeE,
There exists in
{E} (-00 < x < (0)
E,
E
and
i k,
and hence
M IE z
b. 1 < k 2.
dim(E
e Lat Mz
contained in
e
00,
F) = k,
such that
zE
there such
has
has no cyclic vectors.
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
341
This answers questions raised by various authors. Recently Stefan Richter has shown how to lift some of this pathology into the lattice of invariant subspaces of CHAPTER THREE:
H2
of the polydisc.
THE DIRICHLET SHIFT
In this chapter we consider the Dirichlet space
D.
We recall
that from the corollary to Proposition 6, a cyclic vector must be an outer function.
The converse is not valid; in 1952 Carleson
[11] gave examples of outer functions in
Dk , for any integer given in advance, that are not cyclic in D. For a more
k > 1
general result see Theorem 4 below.
First we recall some facts
about the boundary values of functions in In 1913 Fejer proved that if holomorphic and univalent for f
converges uniformly on
Izl
is continuous for
< 1,
Izl = 1.
(since a univalent function is in finite area).
f
D. Izl ~ 1,
then the power series for
Such a function is in D
if and only if
~(~)
D has
Landau pointed out that Fejer's method proved a
more general result:
for any
feD,
the power series converges
at each boundary point where the radial limit exists, and the series converges uniformly on any set where the radial limit exists uniformly (see [32, §13, pp. 65-67]). valid by Abel's theorem:
Of course the converse is
the radial limit exists at any boundary
point where the power series converges.
In fact, the non-tangential
limit exists at all such points (and uniformly on any set where the power series converges uniformly) • Also, the non-tangential limit exists almost everywhere by the Fatou theory.
This result was improved substantially by
Beurling [5] in 1940:
the non-tangential limit exists except for
a boundary set of logarithmic capacity zero.
Salem and Zygmund
[39] in 1946 gave a new proof of this result, and extended it to the
Da
spaces
(0 < a
~
1),
replaced by a related capacity.
with the logarithmic capacity In 1950, Carleson, in his disser-
tation [10, Chap. III, §3], showed that for bounded functions in D a stronger result is true:
the non-tangential limit exists except
perhaps for a set of logarithmic length zero.
A. L. SHIELDS
342
~
In non-tangential convergence one considers angles in with vertex at a boundary point Stolz angle the limit of
(Stolz angles).
For every
is required to exist as
z
J. Kinney [24] in 1963.
feD
wo
~,
He replaces the Stolz angles by a family
tangent to the boundary at
wO'
and making
an arbitrary finite order of contact with the boundary. that if
~
An important new result was obtained by
inside that angle.
of subregions of
f(z)
w.
He shows
then at almost every boundary point the limit
exists inside any such region.
He also gives information about
the capacities of the exceptional sets where the limits do not exist.
Finally in 1982 definitive results were obtained by Nagel,
Rudin, and J. Shapiro [33, Thm. 1], [34, Thm. A]. if
feD
They show that
then the boundary limit exists almost everywhere inside
subregions making exponential order of contact with the boundary. They show the precise connection between the kind of subregions in which approach to the boundary is permitted, and the size of the exceptional set where the limit fails to exist. As regards cyclic vectors, we consider first the two test questions: 1)
If
f ,g e D,
cyclic, then must 2)
If
f
f and
If(z)
1
> Ig(z)
be cyclice? f- l are in
D,
for
1
z e~,
then must
f
and
g
is
be cyclic?
In [9, Cor. 1 to Thm. 1, p. 281] L. Brown and the author show that 1) has an affirmative answer if, in addition, one assumes that g e M(D).
The also raise the following problem (p. 282).
~UESTION 8.
If
g e 0
there exist a sequence IIPngll""
2
const.,
{Pn}
I]PngIID
2
nH""
is cyclic in
0
then does
of poly?omials such that: const., and
(Pn g ) ~ 1
in
~?
An affirmative answer would improve the above partial answer 00
to the first question: (instead of
the result is correct if merely
g e M(D».
g eon H
In the proof of Theorem 2 of [9] an
affirmation answer to Question 8 is obtained under the additional hypothesis that
igl
is Dini continuous on
a~
(and therefore
343
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
g e A,
the disc algebra).
It is not known whether the answer is
still affirmative if merely
g e DnA.
It is easy to see that if h e D
if and only if
f,g e D
nH
oo
f,g e D
h' e B).
nH
oo
then
fg e D (recall
Assume for the moment that
In [9, Proposition 11] it is shown that if
•
cyclic then both
f
and
g
are cyclic.
fg
is
An affirmative answer to
Question 8 would imply the converse result.
More generally one
has the following problem. QUESTION 9.
If
f,g,fg e D,
cyclic if and only if both
f
then is it true that
and
g
fg
is
are cyclic?
As regards the second test question above, we have the following result: if f,f- l e D Hoo then each is cyclic in D.
n
This follows from one of the results mentioned above, since their product is cyclic. Also, it was shown in [48] that if merely f- l e Hoo , and feD, then f is cyclic. However it is unknown f- l
where
must be cyclic in this case.
QUESTION 10.
If
feD,
If I > c > 0
in
~,
then must
f- l
be cyclic? Note that
f- l
must be in
that the derivative is in
D
(one sees this by showing
B).
The following is a combination of [9, Proposition 13] with one of the results above. PROPOSITION 19. a) and
g
is cyclic, then b)
zeros in
If
h
~,
if
If f
g e M(D),
feD,
If I > Igl
in
~,
is cyclic.
is analytic on Re el > 0,
then
(the closed disc) and has no
~
hel
is cyclic and is in
M(D) •
\'1e turn now to another approach to the problem of classifying the cyclic vectors in NOTATION.
D.
Z (f)
CONJECTURE ([9, p. 292]). is an outer function and
feD
is cyclic if and only if f
Z(f) has logarithmic capacity zero.
344
A. L. SHIELDS
If correct this would immediately imply an affirmative answer to our first test question, where is cyclic. question.
If I > Igl
~
in
and
g
It would also imply an affirmative answer to our second Indeed, as noted at the beginning of this chapter,
Beurling showed that if
feD
then
f
has a (finite) radial
limit except for a set of capacity zero. on a set of positive capacity then
l/f
If this limit were zero would have an infinite
radial limit on this set, and hence could not be in
D.
Finally,
if the conjecture is correct it would yield an affirmative answer to Question 9, since the union of two sets of capacity zero has capacity zero.
The conjecture has been proved in one direction.
THEOREM 4 ([9, Theorem 5]).
If
positive logarithmic capacity, then
feD f
and if
Z(f)
has
is not cyclic.
The following results, which we state without proof, are also relevant.
Recall from Chapter One that if
f' e H2
Also, it follows from the Cauchy inequality that lutely converger.t power series and hence PROPOSITION 20.
a)
[8].
If
E C
b)
D,
such that
[9, Theorem 3].
If
has an abso-
a~
is a closed set of feD (\A
that is
Z(f) = E. If
is at most countable, then NOTATION.
f
f e M(D).
f e A.
logarithmic capacity zero, then there is an cyclic in
then
f f
E C a~
is outer,
f' e H2,
is cyclic in
and if
Z(f)
D.
is a Borel set of positive capacity,
let D = E
Here q.e.
1£
e D: lim f(re
:i8
)=0
(rtl), q.e.
in
EL
(quasi-everywhere) means except for a set of capacity
zero. THEOREM 5.
DE
is a closed subspace of
D.
This result was pointed out to us by J. Shapiro.
It is a
consequence of a result in his paper [34] with Nagel and Rudin. See [9, p. 295] for more details.
In 1952 Car1eson [11, p. 335]
345
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
proved this in the special case that Two) of positive capacity.
E
is a BCH set (see Chapter
(Actually, he proved a slightly dif-
ferent but equivalent result.)
Theorem 4 above is an immediate
corollary to Theorem 5, though it is simpler to prove Theorem 4 directly. QUESTION 11.
If
E,PC:d~
are Borel sets, when is
DE
Clearly this will happen if the symmetric difference of and
F
E
has logarithmic capacity zero, but this is not the only
case.
Indeed, Carleson [II, Theorem 5J has given a sufficient
condition for
DE
to equal
{O},
a set of uniqueness for the space Let
[f]
multiples of
in other words, for
E
to be
D.
denote the closure in
D
of the polynomial
f.
QUESTION 12.
E
Dp 7
=
If
feD
is an outer function, is
[f]
DE'
If
MC D
is a (closed) invariant subspace
Z(f)7 QUESTION 13.
(that is,
zM CM),
must there exist
feD
such that
M
[fJ?
=
More generally, one might ask for a chacterization of all invariant subspaces, perhaps along the lines of Korenblum's description [28] of the invariant subspaces of the space (those functions whose first derivative is in cally,
let
and let
E
M(E,~)
d~
be a Borel set, let
~
2
H ).
be an inner function,
denote the set of all those functions in
whose inner factor is divisible by M(E,~)
is easy to show that
~.
2
HI More specifi-
DE
In view of Theorem 5 it
is a closed subspace of
D,
and it
is clearly invariant. 9UESTION 14.
Is every invariant subspace of the form
M(E,~)?
Even if this has an affirmative answer, there will still remain the problem of deciding when two of these subspaces coincide (see Question 11).
Also, this raises the question of describing
those inner functions that can occur as divisors of functions in D.
346
A. L. SHIELDS
One answer to this question was given by Carleson [12] who gave a formula expressing the Dirichlet integral in terms of the canonical factors:
outer, singular inner, and Blaschke product.
However
it still seems very difficult to give a complete description of the zero sets, in
L,
of functions in
D, for example.
Carleson
showed that a Blaschke sequence with just one limit point on the boundary may fail to be a zero set for in [13]).
D
(the proof is presented
On the other hand, there are zero sets such that every
boundary point is a limit point of zeros.
For a "best possible"
sufficient condition that a sequence
{z} be such a zero set see n [45]; the condition is in terms of the moduli {Iz I} alone. n
Finally, we state three more questions about the space If
1¢
L,
in
(z) 1 < 1
9UESTION 15. D?
vllien
C¢
let
For which
a mUltiplier on
a bounded ogerator on
for some
E: > 0, must
¢
be
Is a random power series from the Dirichlet . 1 1 lp "2 ?
More precisely, i f
12
¢.
D?
space almost surely in
< 00,
if
for almost every
t?
n
is
¢ € D nLipE:
If
QUESTION 17.
Znla
¢
0
is bounded must it preserve cyclic vectors?
QUESTION 16.
if
f
D.
{r (tl} n
1
(z) = Zr (t)a zn, then is f t in lip 2 n n See Duren [14, Chap. 5] for information f
t
It follows from [22, Chap. VII, Thm. 2]
about Lipschitz classes. that, for almost every
are the Rademacher functions,
t,
ft
e
Lip a
for all
a <
1 "2.
REFERENCES [1]
Aharonov, A., Shapiro, H.S. and Shields, A.L., "Weakly invertible elements in the space of square-summable holomorphic functions", J. London Math. Soc. (2) 9 (1974/5), pp. 183-192.
[2]
BeauzamY,B.,"Unoperateur sans sous-espace invariant; simplification de l'example de P. Enflo",preprint.
[3]
Bercovici, H., Foias, C., and Pearcy, C.M., "Invariant subspaces, dilation theory, and dual algebras", CBMS regional conference series in mathematics, Amer. Math. Soc., Providence, R.I. , to appear.
347
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
[4]
Berman, R., Brown, L. and Cohn, W., "Cyclic vectors of bounded characteristics in Bergman spaces",preprint.
[5]
Beurling, A., "Ensembles exceptionnels", Acta Math. 72 (1940), pp. 1-13.
[6]
, "On two problems concerning linear transformations in Hilbert space", Acta Math. 81 (1949), pp. 239-255.
[7]
Bourdon, P., "Cyclic Nevanlinna class functions in Bergman spaces", preprint.
[8]
Brown, L. and Cohn, W., "Some examples of cyclic vectors in the Dirichelt space", preprint.
[9]
Brown, L and Shields, A.L., "Cyclic vectors in the Dirichlet space",Trans. Amer. Math. Soc. 285 (1984), pp. 269-304.
[10] Carleson, L., "On a class of meromorphic functions and its associated exceptional sets", Appelbergs Boktryckeri, Uppsala, 1950. [11]
, "Sets of uniqueness for functions regular in the unit circle", Acta Math. 87 (1952), pp. 325-345.
[12]
, "A representation formula for the Dirichlet integral", Math. Zeit. 73 (1960), pp. 190-196.
[13] Caughran,J.,"Two results concerning the zeros of functions with finite Dirichlet integral", Can. J. Math. 21 (1969), pp. 312316. [14] Duren, P., Theory of HP spaces, Pure and Appl. Math., vol. 38, Academic Press, New York (1970). [15] Duren, P., Romberg, B.W. and Shields, A.L., "Linear functionals on It' spaces with 0 < P < 1 ", J. reine angew. Math. 238 (1969), pp. 32-60. [16] Enflo, P., "On the invariant subspace problem in Banach spaces", Acta. Math., to appear. [17] Garnett, J., Bounded analytic functions, Pure and Appl. Math., vol. 96, Academic Press, New York (1981). [18] Hartman, P., and Kershner, R., "The structure of monotone functions", Amer. J. Math. 59 (1937), pp. 809-822. [19] Hayman, W., "On Nevanlinna I s second theorem and extensions", Rend. Circ. Mat. Palrmo (2)2 (1953), pp. 346-392. [20] Hoffman, K., Banach spaces of analytic functions, Hall, Englewood Cliffs, New Jersey (1962).
Prentice
[21] Horowitz, C., "Zeros of functions in Bergman spaces", Duke Math. J. 41 (1974), pp. 693-710. [22] Kahane, J., Some random series of functions, and Co., Lexington, Mass. 1968.
D.C. Heath
348
A. L. SHIELDS
[23]
Kelley, J., General Topology, van Nostrand Co., New York, 1955.
[24]
Kinney, J., "Tangential limits of functions of the class Sa.", Proc. Amer. Math. Soc. 14 (1963), pp. 68-70.
[25]
Koosis, P., Introduction to HP spaces, London Math. Soc., Lecture Note Series 40, Cambridge Univ. Press, London, 1980.
[26]
Kopp, R. P., "A subcollection of algebras in a collection of Banach spaces", Pac. J. Math. 30 (1969), pp. 433-435.
[27]
Korenblum, B.I., "Functions holomorphic in a disc and smooth in its closure",Doklady Akad. Nauk. SSSR 200 (1971), pp. 2427; English translation: Soviet Math. Doklady 12 (1971), pp. 1312-1315.
[28]
, "Invariant subspaces of the shift operator in a weighted Hilbert space", Matern. Sbornik 89 (131) (1972), pp. 110-137; English translation: Math. USSR Sbornik 18 (1972), pp. 111-138.
[29]
, "An extension of the Nevanlinna theory", Acta Math. 135 (1975), pp. 187-219.
[30]
, "A Beurling-type theorem", Acta. Math. 138 (1977), pp. 265-293.
[31]
, "Cyclic elements in some spaces of analytic functions", Bull. Amer. Math. Soc. 5 (1981), pp. 317-318.
[32]
Landau, E., Darstellung und BegrUndung einiger neuerer Ergebnisse der Funktionetheorie, Zweite Auflage, J. Springer Verlag, Berlin, 1929.
[33]
Nagel, A., Rudin, W. and Shapiro, J.H., "Tangential boundary behaviour of harmonic extenstions of LP potentials", Conference on Harmonic Analysis in honor of Antoni Zygmund, vol. II, 1982, pp. 533-548, ed. W. Beckner et aI, Wadsworth Publishers, Belmont, Calif.
[34]
, "Tangential boundary behaviour of functions in Dirichlet-type spaces", Annals of Math. 116 (1982), pp. 331-360.
[35]
Nelson, J.D., "A characterization of zero sets for Michigan Math. J. 18 (1971), pp. 141-147.
[36]
Read, C.J., "A solution to the invariant subspace problem", Bull. London Math. Soc. 16 (1984), 337-401.
[37]
_ _~___ ' "A Rolution to the invariant subspace problem on the space ~l, preprint.
[38]
Roberts, J., "Cyclic inner functions in the Bergman spaces and weak outer functions in HP, 0 < P < 1 ", Illinois J. Math. (to appear) .
00
A "
CYCLIC VECTORS IN BANACH SPACES OF ANALYTIC FUNCTIONS
349
[39]
Salem, R. and Zygmund, A., "Capacity of sets and Fourier series", Trans. Amer. Math. Soc. 59 (1946), pp. 23-41.
[40]
Shamoyan, F.A., "Weak invertibility in some spaces of analytic functions", Akad. Nauk Armyan. SSR Doklady 74 (1982), no. 4, pp. 157-161; MR 84e: 30077.
[41]
Shapiro, B.S., "Weakly invertible elements in certain function spaces, and generators in £1 ", Mich. Math. J. 11 (1964), pp. 161-165.
[42]
, "Weighted polynomial approximation and boundary behaviour of analytic functions", Contemporary Problems in Analytic Functions (Internat. Conference, Erevan, Armenia 1965). "Nauka" Moscow 1966, pp. 326-335.
[43]
, "Some remarks on weighted polynomial approximation of holomorphic functions", Mat. Sbornik 73 (115) (1967), pp. 320-330; English translation: Math. USSR Sb. 2 (1967), pp. 285-294.
[44]
, "Monotone singular functions of high smoothness",Mich. Math. J. 15 (1968), pp. 265-275.
[45]
Shapiro, B.S. and Shields, A.L., "On the zeros of functions with finite Dirichlet integral and some related function spaces", Math. Zeit. 80 (1962), pp. 217-229.
[46]
Shapiro, J.B., "Cyclic inner functions in Bergman spaces", preprint (1980), not for publication.
[47]
Shields, A.L., "Weighted shift operators and analytic function theory", Math. Surveys 13: Topics in operator theory, ed. C.M. Pearcy, Amer. Math. Soc., Providence, R.I. (1974), pp. 49-128 (second printing, with addendum, 1979).
[48]
, "Cyclic vectors in some spaces of analytic functions", Proc. Royal Irish Acad., 74 Sect. A (1974), pp. 293-296.
[49]
, "Cyclic vectors in spaces of analytic functions", Issled. po lin. oper. i teor. funkc., Zapiski naucn. Seminar LOMI 8 (1978), pp. 142-144.
[50]
, "An analogue of a Bardy-Littlewood-Fejer inequality for upper triangular trace class operators", Math. Zeitsch. 182 (1983), pp. 473-484.
[51]
Stegenga, D.A., "Multipliers of the Dirichlet space", Illinois J. Math. 24 (1980), 113-139.
[52]
Taylor, B.A. and Williams, D.L., "Zeros of Lipschitz functions analytic in the unit disc", Michigan Math. J. 18 (1,)71), pp. 129-139.
[53]
Taylor, G.D., "Multipliers on 123 (1966), pp. 229-240.
Det ", Trans. Amer. lftath. Soc.
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
N. J. Young University of Glasgow
1.
Non-uniqueness and one step extensions. There are some classical interpolation problems for complex
functions, solved many times over decades ago, which still generate dozens of new papers. mathematical:
One reason for this is strictly
the development of certain branches of operator
theory has enabled us to view the problems in a different way and so understand some aspects of them better.
There is another
reason, which for me rings an even louder bell:
contact with the
theory of circuits and systems developed by engineers.
An
enor-
mous range of worthwhile problems about complex functions comes from this source.
Many of them are close to questions studied by
the old masters of analysis:
for example, variants of the
Nevanlinna-Pick interpolation problem arise from a remarkable diversity of starting points. back to
19~O
The earliest instance I know dates
(see the account of F. Fenyves' work in (9) 3S1
s. C Power fed.}, Operators and Function Theory, 351-383. c> 1985 by D. Reidel Publishing Company.
while,
N.J. YOUNG
352
more recently, J.W. Helton's far-reaching application of nonEuclidean functional analysis to electronics also centres around this problem [4].
However, the engineering slant generally calls
for something slightly different from the old results.
In the
first place, one is generally aiming eventually at a method for the practical computation of solutions, and so numerical considerations such as stability, conditioning and storage requirements playa role.
A second major difference is that in most applic-
ations the functions one has to deal with are matrix-valued rather than scalar.
This is essentially because the states of most
interesting engineering systems are described by vectors rather than scalars:
solving a system of linear differential equations
with constant coefficients using the Laplace transform introduces a matrix of rational functions.
Now numerical complex analysis
and the theory of analytic matrix functions are both substantial mathematical topics which have received plenty of attention from mathematicians independently of any practical implications.
Still.
this contact with engineering deserves to be heartily welcomed by the pure mathematicians in the field.
It is surely desirable
that our subject be in contact with other branches of science and, more concretely. it provides an orientation and a public in an otherwise large and diffuse area. In these lectures I shall show how the two factors I have mentioned - matrix valued functions and the search for efficient numerical algorithms - affect the Nevanlinna-Pick interpolation
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
problem.
353
Let us begin with a modification of the versions pre-
sented in the lectures of D. Sarason.
Let
be
distinct points in the open unit disc D and let Wl ' complex numbers.
.(z.) = 1.
w., 1.
in D satisfying the interpolation
• 1 ~
i
to belong to the unit ball of functions in
D
~
n,
r
but instead of asking for -. (the space of bounded analytic
••
show using nOl'llal families that the infimum of
f
11.11 Hco
with supremum norm). we ask that
minimised over all interpolating functions
ed.
be
It is easy to
11.11
Bco
is attain-
We shall re-formulate this problem somewhat. Suppose that aD is any B function satisfying the interpolation conditions -
for example, the Lagrange interpolating polynomial. second function
• - f
I
is divisible by the polynomial
- 8.8
J
Then a
• E HaD satisfies those conditions if and only if
or in other words, if and only i f As
be
As in Sarason I s problem (A) we look for a
bounded analytic function conditions
••• , •.• wn
• E
(z -zl)(z -z2) ... (z - 3 n ),
f +
is a unit in the ring BaD, (z - 8
1
)
...
where (as in [8]) b(z)
=
n
TT
= bB aD,
z - z.
j=l 1 -
the finite Blaschke product with zeros arrive at the following.
we have
(z - 8 )B-
n
(z -8n )B-.
(8 -zl)
i1
,
8.3
J
8 1,
... ,
8
n'
We thus
354
N. J. YOUNG
Problem (C). find
Given f € Hoo and a finite Blaschke product
• € f ... bH oo ~ ~
11.11 H-
b,
is minimised.
As was indicated in Sarason's lectures, there are numerous approaches to this problem which, from a purely mathematical viewpoint, seem to be of roughly equal power.
For computation,
though, some are more convenient than others.
The Krein space
method gives the theoretical results in a most elegant fashion, but its prescription for obtaining the extremal
•
contains the
step "Pick a maximal negative z-invariant subspace of containing the negative subspace ••• ".
H2
®
H2
While one could doubtless
develop methods of handling such injunctions on a computer, for the present a more straightforward option is to use an approach based on familiar, concrete entities like singular values of matrices.
For this reason my own computer programs [2] for the
numerical solution of Problem (C) do not use the methods modestlypresented by the earlier speaker, but rather Sarason's own operator-theoretic approach [7].
Here is his construction.
Corresponding to the data in problem (C) above we define an operator
from the Hardy space
T
H2
to its subspace
g2
e
bH2
by (1.1)
where and
Mf
P :
Note that
:
H2 _H2
_ 82
H2
T
8
is the operation of multiplication by f
bH2
is the orthogonal projection operator.
depends onJ.y on the coset
f ...
bHa:
not on
t
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
itself, for if
.:
355
f + bh is a typical element of this coset,
z E
bhz E bH2
P(bb): 0
and
Sarason proved that the opposite inequality is also true:
the
we have, for any
H2,
so that
hence
It follows that, for any
• E
f + bH flO ,
lip 1/
:
<
Hence
IITII
proof is a good deal less trivial than the foregoing. ingredients are a duality argument and the fact that an function can be factorized into the product of two This gives us the function
•
H2
The main
HI functions.
He norm of the desired extremal interpolating
as the operator norm of
T.
We can also obtain
•
itself from the analysis of T •
Theo'l'em 1 [7]
The unique sptution •
of Pztob1.em (C) is
• : Tv/v whezoe T is the opemtOl' defined in vector of T.
MOl'eovel'.
11.11 B-
(1.1)
:
and v is any mt:IZ'lmising
\I Til.
By a mazimising veato.r for T we mean a non-zero vector
356
N.J. YOUNG
T is a finite
as
rank operator, there do exist maximising vectors.
Once the
final norm equality is established the formula for
+
easily:
taking
V
to be a unit vector, we have
II p( +v ) II
It follows that
=
!I 'v II,
,.
As
v we obtain a formula
IITII
singular value
•
corresponding to the
(or in other words, an eigenvector of
nTIi· ~),
corresponding to the eigenvalue mination of
T
is a singular vector for
V
+v = p( +v) = Tv.
and hence
On dividing through by the scalar function for
follows
T*T
this reduces the deter-
to a singular value problem for a finite rank
operator.
Nevanlinna-Pick for matrices. In the work of Helton, among others, one encounters the most natural possible generalization of Problem (C) to matrix-valued functions.
Let us
...mxn
for the space of m x n
write~·
complex
matrices with the operator norm:
HAil where
~
and
~n
=
sup{
11.4=11
0:
m
1I.x11 c n -<
l},
have the usual Euclidean norm (numerical
analysts call this the speotral norm of A). the space of bounded analytic
-
Hmxn
will denote
~n-valued functions on
D with
357
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
supremum norm: sup sE D
F in turn shows that
Fatou's theorem applied to each entry of F E Boo
has a radial limit almost everywhere on
mxn
L~
is naturally embedded in the space urab le
a!"xn -valued
.
funct~ons
essential supremum norm. B(s)
inner i f
on
3D,
A function
so that
aD,
mxn of bounded measmodulo equality a.e., with
B E Boo
mxm
is said to be
is a unitary matrix for almost all
s E 3D.
One
can show quite easily by dividing out zeros that the general
m
rational inner function of type B(z)
where k
::
U1
l~(a)
;J
x
r ;J
m has the form
U2 ••• Uk
is a non-negative integer, each
unitary matrix and each
b .(z) J
~ (a)
U. J
Uk +l
is a constant
is a scalar Blaschke factor.
Such an inner function is called a Potapov-BZaschke product. For present purposes it will do no harm to think of all inner matrix functions which we encounter as being of this type. ProbZem (D).
Given
find
t
and inner fUnctions
in the coset
F + BBoo
BE
mxn C such that
rmxm lit II
00
is minimised.
Once again, consideration of normal families shows that the infimum of
\I til
00
is attained.
To simplify the exposition we
358
N. J. YOUNG
shall take C to be the constant function equal to the identity matrix at every point.
Mathematically, this involves no loss of
generality, as problem (D) is equivalent to minimising the norm over the coset F(adj C) + B(det C) H;:n' where
(adj C)(z)
is the adjugate matrix of
ally, however, the use of
C(z).
Computat ion-
adj C is best avoided.
What is the analogue, for Problem (D) (with C a I) of Sarason's construction?
The multiplication operator Mf : 8 2
eralizes to a multiplication operator M
-+
'H2_H2 F' n m
= Here
'
82
gen-
given by
FCz) x(z).
denotes the Hardy space of ~-valued analytic functions
in the disc, which we may think of as the space of n x 1 column vectors of scalar
H2
functions, with the obvious inner product.
Sarason's operator thus generalizes to
T = PMF'• 8n2 where operator.
G
BH 2 m
_
82
m
G
BH2 m
(1.2)
is the orthogonal projection
As an initial guess we might hope that (at least for
rational inner functions) the solution of Problem (D) would be unique and that the extremal result as the following.
t
could be computed from some such
359
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
Let F € H~n and Zet B be an inner jUnction in
Theorem 2.
Let ~ be an eZement of minimaZ norm in the coset
BOO. mxm F
in
+ BB;n'
II ~ II
Then
00
FUrthermore~
(1.2).
=
II TII ~
bJhere T is the operator defined T~
if v is a maximising vector for
tv = Tv. A reference for this result in the above form is [11];
how-
ever, equivalent statements are known to all authors who have published on the Nevanlinna-Pick problem for matrices - see for example [3). A second look at the theorem will reveal that it guarantees us much less than our initial guess.
There is no uniqueness
statement, and if we try to calculate
t
tv
=
from the relation
Tv, has been found, we
assuming that a maximising vector
shall find that we have insufficient data ( in the case n > 1).
v is a column vector, so we cannot divide through to get t. it happened that the relation
tv
V
= Tv
were of the form
[vI 0 ••• 0],
If
VI € ~,
t:
would give us only the first column of
in general, it roughly speaking determines a rank 1 part of
t.
How then are we to get further equations from which to determine t ?
The answer is that without further inputs (e.g. the
imposition of further conditions on itrary choice)
we cannot determine
• t,
or the making of an arbbecause the solution of
problem (D) is in most cases very far from being unique.
A
360
N. J. YOUNG
simple example will make this clearer. Let
m=n =2
diag{2, l}.
G
e:
F
e:
Let
of all funct ions IIG(O} II
and let
= 2, 00
be the constant function F + BB oo
e:
consists
2x2
such that G(O} = diag{2. l}. Since 2x2 the maximum principle tells us that II Gil 00' for G
Boo
On the other hand, F
of norm 2, and so the infimum of precisely 2.
2X2
The coset
is at least 2.
F + BB2x 2.'
Boo
For which
t
IItll
00
for
t e:
F + BB oo
is
2x2
is this norm attained?
maximum principle shows that the (1, l) entry of
is itself
t
The (scalar) must be con-
stant and equal to 2, and hence the two off-diagonal entries must be identically zero.
Any extremal
t
must therefore be of the
form
where
cjl22(0}
=1
is an element of example is very
and
II cjI" 22 HP:> 00
F + BB2.x2 instructive~
< 2; conversely, any such
-
of minimal norm.
This trivial
it shows that the supremum norm is
too weak an indicator of the "size" of a matrix-valued function for us to expect any useful uniqueness results in problem (D). One natural response to the profusion of solutions to problem (D) is to describe all of them.
This was done (in
th~
equivalent block Hankel matrix formulation) by V.M. Adamyan, D.Z. Arov and M.G. Krein in [1].
Let us examine their strategy
and see how it can be used to compute a solution of problem (D). This reformulation of the relevant part of (1] is taken from
361
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
F. B. Yeh's thesis [10].
For simplicity we take
We start from the observation that if problem (D) and vI' v 2
Hence, if B2 2
V
t
m: n
= 2.
is an extremal for
is any maximising vector for
T then
tv
= Tv.
are linearly independent maximising vectors in
we have the following relation between 2 )( 2 matrices:
=
t(s)
(1.3)
Actually, we need a stronger property than linear independence for
V
I
and
V •
2'
we require them to be
ent, in the sense that 4: 2
for almost all
vIes), V2 (s)
e: aD.
s
are linearly independent in
The vectors 8 22 ,
are linearly independent in independent.
pointtuise UnearZy independ-
[1
s]T and
[.3
s2]T
but not pointwise linearly
If we can choose the maximising vectors with this
stronger independence property then we can solve (1.3) to obtain
a.e. on aD, so that
•
is uniquely determined in this case.
Since we already know that
•
is not in general unique, on the
face of it it appears that all we can conclude is that there do not always exist pointwise linearly independent maximising vectors for
T.
Nevertheless, Adamyan et aZ. contrive to reduce the
general case to the special one in which such a exist. B
They effect this by modifying the given functions
in a subtle manner.
function
VI and V 2
F:
F + BFo
Observe that, for any lies in the coset
Foe: t2 x2 ,
do
F and the
F + B8;:2' and hence
N. J. YOUNG
362
-
, + zBEfO
2x2
c:
=
-F
+ BBI» 2x2
F
+ BBI»
2X2
11.11 I» over the left hand coset is thus no less
The infimum of
than its infimum over the right hand coset, the quantity called for in problem (D).
If we can choose
Fa
so that the two
infima coincide then any extremal function in the left hand coset will be a solution of problem (D).
And as the left hand coset
is smaller it can in principle contain fewer functions of minimal norm. which
F + zBB CIO 2X2
P
for
contains a unique function of minimal norm,
and moreover, the Sara son operator corresponding to
where
Fa
Adamyan et at. show that there is a choice of
B2 GaBB2 222
8 2 _
this coset,
is the orthogonal projection,
does admit two pointwise linearly independent maximising vectors. Hence if we can find
Fa E
¢2X2
(necessarily non-unique in
general) such that and
(1)
(2)
T has
two pointwise linearly independent maximising vectors,
then we can use the idea outlined above to compute a solution of problem CD).
The operator
T
is a "one step extension" of
T:
T
in terms of block Hankel operators, passing from
T
corresponds to the addition of an extra column.
I f this lIethod
to
is applied to the example given above it entails passing from
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
363
the coset
+
sH
co
2x2
to anyone of the smaller cosets
with
IAI:
The latter coset contains a unique function of
1.
minimal norm, to wit
[: It is interesting that no choice of "natural" solution Algorithms
A (i.e. of
Fo) gives us the
t . diag{2, I}.
for the solution of problem (D) or its equivalent
by the method of one step extensions have been implemented by S.Y. Kung and D. Lin [5] and by F.B. Yeh [10].
Fo
sired matrix find one.
Although the de-
is known to exist, it is no trivial matter to
Kung and Lin achieve it by solving an algebraic
Riccati equation for matrices: ive procedure.
this involves at least one iterat-
Yeh's method is essentially rational, but is still
a substantial numerical step, and there is evidence that it may be rather unstable.
Both methods must inescapably require the pro-
gram to make an arbitrary choice at some point.
364
N. J. YOUNG
Here is the method used in Yeh's implementation for finding the desired one step extension. and (2) above are satisfied: such an
Po
Let
F
o
E
be such that (1)
by the result of Adamyan et aZ ••
The codomain of
does exist.
a:2 x2
T
admits the orthog-
onal decomposition ::
consists of the constant functions in identify this subspace with gonal projection of its codomain
~t
and denote by
Po
H2. 2 •
let us
the ortho-
With respect to this decomposition
-
T can be written as a block operator matrix
::
where the multiplication operator here is regarded as acting from
Hl
Ll.
into
II Til::
p
B*F + Fo
and let
V
being non-analytic in general.
be a maximising vector for T (there are
such vectors in the rational case or. more generally, when is the sum of a continuous and an compact) •
Let
The requirement that
B*F
H oo function, as T is then
II Til:: IITII
clearly entails
that
::
(1.4)
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
to be satisfied by
FO.
365
The second requirement is that
pointwise linearly independent maximising vectors. the pointwise orthogonal complement of
V in
Let
82
p
exactly.
V
be
We require
2
to have norm
T have
Thus, if we write
we require that
and furthermoroe, the left hand side is singular. operator
Q : V -
8 2 <:) B8 2 2
2
The condition is then Q*Q
such that
~R*R
operator
such that X
must satisfy, for any
= g
e: V n
B8
=1
and
R
= XQ.
g e: V,
= PO(B*F+FO)g
XQg
Hence, i f
, the difference between the two
This is equivalent to the existence of an
sides being singular.
That is,
Construct an
2,
2
XQg
Po(B*Fg) + Fog(O>.
so that
=
g(O>
(1.5)
= 0, we have (1.6 )
N.J. YOUNG
366
On the other hand, if
g E V and g(O)
= v(O)
we have, from
(1.5) and (1.4),
= Po (B*Fgl + Fog(O)
x Qg
=
(1. 7)
Po(B*Fg) + Po(B*PV).
Conversely, one can show that if X : H22
G
BHl
.. c2
(1.6) and (1.7) and has norm 1 then the operator R form
PoMB*F +F
o
IV
for some
In the rational case
H22
8
Fo BHl
= XQ
satisfies has the
having the desired properties. has finite dimension and the
relations (1.6) and (1.7) reduce to a finite set of linear constraints
where the vectors
Zj' Vj can be computed.
Finding X of norm 1
(when there is one) satisfying these conditions is a straightforward piece of linear algebra..
X's:
it is at this point that the arbitrary choice is made in
Q is also a known operator, and hence
Yeh's implementation.
R
=XQ
FO'
Typically there will be many such
can be calculated.
From this point it is not hard to find
Library routines can be used to find a pair of independent
P,
maximising vectors for struction of function
•
and formula (1.3) completes the conof minimal norm in
F + BH;:2.
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
2.
367
Strengthened minimisation and the restoration of uniqueness. In the last lecture I dropped a few hints which were intended
to suggest that the results of Adamyan et
at~,
and successors who
have used one step extension techniques, are not the last word on the matrix Nevanlinna-Pick problem, particularly for those interested in the computation of solutions. the specialization
Given t E F
F E H
CD
m~
+ BHmn CD X
Recall problem (D), with
C. I: and an innel' function
such that
II til GO
B E H CD
m~
,
find
is minimised.
The procedure outlined above involves constructing ....
FEF1'BH
CD
mxn
unique 1IIember
such that the smaller coset t
F + aBH CD contains a mxn
-
of minimal norm, and that this
of the original problem.
is a solution
The most concrete objection to this
method is that the calculation of numerical stability.
t
F is lengthy and of uncertain
However, my main reason for putting forward
an alternative approach is rather mathematical intuition. is not inconsistent with having an eye to applications.
This Esthetic
considerations have long been inexplicably effective in theoretical physics, and I have faith that the same will be true in engineering. See the final two pages of [4] for poetic thoughts on this topic. A less tangible objection to the method of one step extensions is the untidiness of making an arbitrary choice, which is necessitated by the non-uniqueness of the solution of problem (D).
368
N. J. YOUNG
Another objection is that the solutions obtained by this method do not seem to be fully consonant with the spirit of the Nevanlinna-Pick problem.
To explain what I mean, let me return to
the earlier example: m
= n = 2,
F
= diag{2,
l},
B
= sI.
The solutions of problem (D) in this case are all functions of the form
• =
diag{2, g}
IIgIIH .... s.2.
where g(O) = land
The solutions obtainable by
the method of one step extensions satisfy on
aD.
t
identically
U.II ....
equal to values of
.(3)
aD.
almost everywhere on
.(3)
are constant and That is, the singular
are actually as Large as they can be, consistently
being a solution of problem (D).
Nevanlinna-Pick problem as being to minimise available sense, over a coset ask for solutions minimised.
any
of problem (D) obtained in this way is such that all
the singular values (or a-numbers) of
•
=2
This illustrates a general property of the method:
solution
with
Ig(s)1
•
F
+ BB
If we think of the
.,
in the strongest
....
mxn ,then it seems natural to
for which all singular values of
.(3)
are
This is practically the opposite of the one step ex-
tension approach.
To make the formulation precise, let us write 8
0 (A) ~ Sl(A) ~
for the singular values (eigenvalues of
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
and, for
369
write
~
st(F)
sup B€D
;j
8 ,(Fea»),
~
0, 1, 2, , •••
J
Let
The first term of this infinite sequence of non-negative numbers coincides with
UFIICD ,
and at most
I shall call the following the
min{m, n}
teI'llls are non-zero.
strcngthened NevanLinna-Pick
probLem.
Probtem
Given
(E).
F€
BCD
mxn
and inner functions
C € BCD ,. find t € F + BBCD C such that nxn mxn lA1ith respect to the Le3:icographic ordering.
soo(t)
B€
BCD
mXm ' is minimised
Again an argument using normal families shows that the infimum of
SCD(t)
is attained.
Let us look at problem (E) for our
2 x 2
example.
any solution (2.1) of problem (D) we have, for any and hence
::
solve problem (E), we need to minimise g(O) :: 1.
g • 1,
B
€
For
D,
:: IIgll em' Thus, to II gil BCD subject to
By the maximum principle, this has the unique solution
and hence problem (E) has the unique solution
• = diag{ 2,
l}.
In this case the strengthening of the lIinim-
isation condition brings about a unique solution, and the "right" one at that.
N. J. YOUNG
370
The example is typical: ution whenever CD
H
problem (El does have a unique sol-
B*FC* is the sum of a continuous function and an
function, and so in particular in the rational case [11].
I infer that the strengthened Nevanlinna-Pick problem (E) is a sensible generalization to matrix functions of the scalar problem (C). It remains to solve this generalization.
We wish to find
an algorithm capable of efficient numerical implementation, generalizing the one based on Sarason's treatment of the scalar problem.
In considering problem
duced the analogue
CD}
above we have already intro-
T of Sarason's operator. T certainly contains T
all the data, so it is plausible to hope that an analysis of will yield the desired extremal function
t.
The proof that this
is indeed so is lengthy and technical - even the statement of the algorithm in the general case makes it sound complicated.
I shall
describe a special case in an attempt to show that the ideas are quite natural.
Consider a rational
inner function
CD F € H2x2
We shall study the strengthened
Nevanlinna-Pick problem for the coset shall look for the element 8;-<1»
and a rational
t
that is, we
of this coset such that
is lexicographically minimised.
(8;( t).
Form the Sara son
operator
where P : B 22 _
H 2 1':\ BB 2 2 \7 2
is the orthogonal projection.
is compact and therefore admits a unit maximising vector
"o€
T B 22 •
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
371
As we have already seen, one relati9n for
t .is
(2.2 )
F + B8;:2 of minimal
Indeed, this holds for any element of norm.
Let us denote this minimal norm by
further relation for
toe -=
IITII).
A Bl~(t).
will come from consideration of
t
Vo
To make the idea clear, suppose that
and Tv
o
are of an
especially simple form:
= for some a, b € 02. column of
-=
[:J ·
Relation (2.2) then gives us the first
t:
= Now it is simple to show that Ib(s)1 that
b/a
II t II
CD
a.e. on
The (I, 1) entry of
is an inner function.
has modulus equal to
= la(s)1
a.e. on
3D,
the (I, 2) entry must be identically zero.
3D,
so thus
t
and this implies that We deduce that
t
is of the form
(2.3)
for any
It22 (s)1
t
CIO
€ F + BB2x2 for
s £ 3D,
of minimal norm. and so
s:-(t)
=
11.22 11
BOI>.
Thus
81 (.(s»
=
N. J. YOUNG
372
Minimising
CD
•
81
1S
over all possible
thus equivalent to minimising the .22's;
it turns out that this is a scalar
Nevanlinna-Pick problem. is of the fonn CD
Suppose that the second column of
F2~T
[0
F
we may do this since the coset
..
always contains one function of such a type, to wit
F ... BH2x2
The set of possible second columns for
• EFT BH2X2
••
of the form
(2.3) is then, in a self-explanatory notation,
As all w*-closed a-invariant subspaces of \)H CD for some inner
\),
it follows that the (2, 2) entries of
elements of minimal norm in elements of norm at most remal
•
norm in
to
+
F
in
is obtained by taking F22 '" \)H-:
are of the form
8CD
CD
comprise precisely those
BH2x2 F22
... \)0'.,.,
·22
to be the element of minimal
and the desired ext-
We could compute a suitable
F22
and
\)
and solve the resulting scalar Nevanlinna-Pick problem by an established method, but it would be far more satisfactory to exWe can do this from the operator T already computed. ... since the Sarason operator T corresponding to the coset
tract
.22
F22 ... \)H CD
is just a part of T.
In fact, if
wise orthogonal complement of Va in
82.2
in this special case). it turns out that missing entry of
•
VI
denotes the point-
= {o} 0 ® T = T Iv1.
(Le. VI 0
82
The
can therefore be obtained from a maximising Indeed, we have
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
373
= where
(2.4)
are pointwise orthogonal on
aD.
Since
= we can solve (2.4) to give t
=
rvO Tv,] [IIVo:.) 11-2 1
=
I
[VO
OV ,;.) U-2 ]
VIJ*
V.*
(Tv.)
3. 3. IIv.(.)1/2
j=O
J
In this way we obtain an explicit formula for the extremal function in· terms of the maximising vectors of a succession of compressions of the Sarason operator
T-
as nice a generalization of the scalar
result as one could hope for. posed that
Vo and Tvo
In the outline above we have sup-
were of a special form:
the general case
can be reduced to this special case by a diagonalization technique. Theorem 2 of [11]. restricted to 2 x 2 matrices. states the following.
TheOZ'Bm 2. B# C
t
innel'.
101' which
ThB cosilt
s-( t)
(Jl'Q.phic ozod#l'ing.
constzouction.
F
+ BB'z.:2 C contains a unique e"Lement
is minilllised with l'esp8ct to the "Lezicot
can be ca"LcuZated by thB 10"L"Lt:.61i.ng
N.J. YOUNG
374
Let
lh1'e MF : C*a 2 2 and P : L22 -
L22
-
L2
e
BH22
2
is the opezoation of mu.1,tipZication by F is the o1'thogonaZ p1'ojection opeztatozo.
Let v 0
be a unit ma:r:imising vecto1' fo1'
To.
Let
be the oztthogonaZ projection of
c*al
VI
pointwise o1'thogonaZ compZement of Vo in L; the pointwise O1'thogonaZ complement of Tvo Let VI
PI
L; -
opezoat01'.
Then
..
PI MF ; VI -
L;
e
El
1
t
= I j=O
a.e. on
IIT.II J
and "Let El
in 8822
be
•
be a unit ma:r:imising vecto1' fo1' the opezeat01' Tl
WheN
onto the
aD.
a.e. on
L;
e
El
is the o1'thogonaZ p1'ojection
(T. v.) v.te J J J 2 IIv.(·)1I J
Fuztthel'fllON. S/t(3» is constant and equaZ to
aD.
so that
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
3.
375
Comput ing the extreJlal funct ion. The theorem stated in the last lecture in principle tells us
how to obtain the unique solution of the strengthened NevanlinnaPick problem with rational data.
However. quite a lot of math-
ematics remains to be done before we can start writing a computer programme to carry out the calculations.
Hi
ising vectors in
We have to find maxim-
for the Sarason operators
T .• J
and this
entails finding the matrices of these operators with respect to orthonormal bases.
To choose bases which enable us to perform
such tasks efficiently requires technical facts about the spaces of vector-valued rational functions which arise.
This lecture
will be devoted to the algebraic calculations which must precede numerical implementation. Let us begin with the scalar problem (C).
f E Hoo and a finite Blaschke product b.
We are given
oo To find, E f + bB
of minimal norm we form the,Sarason operator
as in (1.1).
T by a matrix the most straightforward
To express
path from a mathematical point of view is to use the natural orthonormal basis the codomain.
1.
2. 22 ••••
for both domain and (a supers pace of)
This results in the Hankel matrix formulation. as
in the papers of Adamyan
et at.
For computational purposes
it is clearly unsatisfactory to use infinite matrices, so we first reduce the problem to an entirely finite-dimensional one.
Note
N. J. YOUNG
376
that any maximising vector for T
is orthogonal to Ker T,
lIIay regard T
Tr
3:
€ 82,
as acting from (Ker
T(b3:):: P(j'b:&)
= 0, so that
to
G
Now 8 2
bB2.
functions.
bH2
g/p
... p
then where
g2
g is
For any
It will
T: 8 2 ~
bB2
-+
is a well known space of rational
Let us denote the numerator of b
denominator by of the form
G
bB2.
bH2 S. Ker T.
therefore suffice to obtain a matrix for
82
G
H2
so we
G
bB2
by
P and its
consists of all functions
...
a polynomial of smaller degree than p.
Everyman's basis for this space, in the case that the zeros of
b are all simple, consists of the functions k.(s) J
( of. Sarason 's Lecture 3).
Sj
of
= (1-3.S)-1 J
This basis is not orthogonal, and so
one must orthonormalise by making use of the Gram matrix
This route leads to the "Pick matrix" - a classical method of solution. antages. b
As a computational method it has two serious disadvIn the first place, it only applies when the zeros of
are simple, and this means that one must expect ill-conditioning
when some of the zeros are close together. actually know the zeros of
to use the Pick matrix.
b
applications we have in mind
Secondly, one must
b
In the
is known as a rational function
expressed as a ratio of polynomials which are not factorized. Although we could simply factorize the numerator over C using a standard library routine, this would introduce error and is plainly a step to be avoided if possible; still greater force
this consideration applies with
to the matrix probl•• (E).
Fortunately there
377
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
a2
is a third basis of need.
I.
Let
~ b~ which has just the properties we
be the degree of
k
be the unique element of
J
of the form
•
p
and, for 0
a2 <:)
a J + O(a ).
;j
~
k - 1, let
having Maclaurin series
bfil
k
~
10 , 11' .•• , Ik-l of
The basis
is not ol!'t:honormal, but it is readily orthonormalised ~
virtue of a very simple expression for its Gram matrix:
where Sk
is the k-dimensional forward shift matrix,
Sk
=
o
0
1
0 0
o
0
o
1
o
0
o
0
1
0
o
(see [2, page 137] for a proof).
We can therefore apply the
Gram-Schmidt process to obtain an orthononnal basis for
1P
GbH2
which is expressed in terms of the coefficients, not the zeros, of
b, and which remains equally applicable when b has very close or even multiple zeros.
Hote that the
I.'s J
are easy to write down.
If
pea) then
pes)
k = a.+a k -1 S+ ••• +a 0 a . I(
There is a polynomial g;j
I. = g ./p. J
h. E J
J
r
of degree less than k such that
By the defining property of the
such that
I.'s, there exists J
378
N.J. YOUNG
..
p
That is, g. (a) J
Hence
g .(a)
= "..,je-ak ...r a k-l"., + •••
=
J
-a.,j + -
ak_l a
k"
The basis obtained from the
j+l + ••• + a. a k-l J+l
f.'s
by orthonormalization cannot
J
be written down so easily, but it can be handled in a computer
without any difficulty. the problem at hand:
It has one further property relevant to
the matrix of Sarason's operator
respect to it can be computed efficiently.
with
It suffices to find
T with respect to the f.'s:
the matrix of
T
the matrix with
J
respect to the orthonormalization of the a similarity transformation.
f.'s is then obtained by J As is customary we denote by 8
the forward shift operator on
H2
shift operator. and hence
T*
T
is the compression of
T*
~(B)
= f(8)-
f(8)*.
8* H2
is the .backward
G
bH2 of f(8),
In fact
JJ2
G
bJJ2
and so we have
=
f(8)*IH2
G
bH2
=
f*(S* IH2
e
bH2)
and the vertical bar denotes restriction.
To get the matrix of T*
fit
so that
is the compression to
is invariant lDlder 8 *,
where
,
E H- to the matrix of
we can thus apply the scalar flDlction S* 1 H2
shows that, with respect to the
Q bJJ2.
fis,
A little calculation
the matrix is a companion
379
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
C
1
0
0
0
1
0
0
0
-aOlak
-a1/ak
-a2 lak
=
•
The special form of efficient.
o o
0
C makes the calculation of !*(C) particularly
We can therefore derive a satisfactory matrix rep-
resentation of T,
after which the computation of a maximising
vector can be effected by a library routine for the singular value decomposition of a matrix.
Fuller details and a report on the
testing of the resulting algorithm can be found in [2]. If we turn to the numerical solution of the Nevanlinna-Pick problem for matrix-valued functions we find many similarities but some substantial differences. special case
C B I.
F € Rco
rmlCm
B €
mxn' in Lecture 2
we
Let us consider problem (E) in the
The data are thus rational functions with Binner.
To apply the theorem stated
must represent the operator
T If
B is a Potapov-Blaschke product with k elementary factors
then the codomain is a k-dimensional space. case we can make
As in the scalar
T into a finite dimensional operator by restrict-
ing it to (Ker T)i, or some superspace thereof. then
is a scalar inner function, and, for any
Let :r;t.
S
=det B:
£ 8 2
n '
380
N.J. YOUNG
= P(Pj3z) = P
T
Thus I3H2 s;. Ker T,
n
T as acting on Hn2
~ ~;,
copies of the k-dimensional space The codomain B 2 G BH 2
already known to us. delicate handling.
O.
and so
We may therefore regard direct sum of n
=
B)Pz)
m
m
which is the 8 2 ~ 1382
requires more
Presumably one could simply use Fourier co-
efficients and thereby reduce to a problem about infinite block Hankel matrices, as in the Kung-Lin implementation of one step extensions.
As far as I know no one has tried strengthened mini-
misation using block Hankel matrices. more promising to generalize the for the scalar problem. isation:
f·J
For computation it looks basis, which works so well
F.B. Yeh [10] has given such a general-
the main ideas are as follows.
To obtain an analogue of the expression element of H2 ~ bg2 ,
g/p for the typical
we must write the rational matrix B
in
fractional form:
B(a)
=
where the "numerator" 8(a)
8(2) D(a)-l and "denominator" D(a)
are matrix
polynomials, and are chosen to have no common right divisors, other thaD the inevitable ones - 1Ulits of the ring are the eleaents with constant non-zero determinant.
which Write
381
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
and let
=
I(a)
A*-tA*18-t
q
The analogue of the!. 's J ion
q-
is the non-orthogonal direct decOilposit-
where F. J
= {'(8)-1
(A q *a j ... Aq-l * a jH -t ••• -t Aj-tl *'aq-l)z:
which is valid provided N(a) is monic
z E..Tn} «.,
~ 0). The q expression in chain brackets gives us a natural identification of
F. J
a!",
with
(i.e. det A
and hence provides a natural basis for
F •• J
Putting these bases together we get a non-orthogonal basis for 8 2
m
aQ
B82
The key ingredient for this approach is that, as in
m·
the scalar case, there is a neat formula for the corresponding Gram matrix, so that the basis can be efficienly orthonormalised. The Gram matrix is
(Imq where
-
B(S) q B(Sq >*)-1
S
B(Sq)
is the q-dimensional shift matrix, as previously, and q denotes the m )( m block matrix whose (1" j)-block is the
q)( q
matrix
b •• (S), l.J q
where
B
=
[b....,]:
.,,,
see (10, S3.3].
It is now clear that we can construct orthon01'lllal bases for H20 88 2
n
Q
n
and
8 2
m
e
BB2
m '
the domain and codomain of T.
It remains to find the _trix of T, and this requires soae rather
N. J. YOUNG
382
intricate calculations.
We can write
T as a composition
is the Sarason operator corresponding to the coset
where
+ SHa)
and IT is the orthogonal projection. TS can be remxn presented by an m x n block matrix in which each block is a scalar
F
Sarason operator, so that the methods of the scalar case apply. Finding the matrix of IT occupies Chapter 4 of [10):
it takes some
weighty manipulations of large block matrices. Computational experience of Sarason operators in the matrix case is still very limited, but what little there is is promising. Storage requirements are rather large, but this seems to be an inevitable consequence of the quantities of data contained in rational matrices.
If det B has degree
dimension nk
k
to one of dimension
then T acts from a space of
k,
l-
so that we must store nk
complex numbers to represent T , and still larger numbers in the intermediate steps.
Although this takes up a lot of room, it is
still less than one needs to work directly with an infinite block Hankel matrix. I hope that the foregoing has given some idea of the mixture of operator theory, function theory and linear algebra which goes into the computation of solutions of modern variants of classical interpolation problems.
I have presented the theory of a single
problem, but the considerations ·and technicalities are typical of many others in the field:
see [6].
INTERPOLATION BY ANALYTIC MATRIX FUNCTIONS
383
REFERENCES 1.
V.M., Arov, D.Z. and Krein, M.G., Infinite Hankel pzoobtems~ Amer. Math. Soc. Trans 1. (2) III (1978) pp. 133-156. Ada~an,
block matzeices and zoelated e:tension 2.
Allison, A.C. and Young, N.J., Nwnezeicat atgozeithms fozo the pzoobtem~ Numer. Math. 42 (1983) pp. 125-145.
Nevanlinna-Pick 3.
4.
Foias, C., Contzoactive intezotb1ining dilations and ",aves in tayezoed media~ Proceedings of the International Congress of Mathematicians, Helsinki (1978). Helton, J .W., Non-euclidean functionat anatysis and Bull. Amer. Math. Soc. 7 (1982) pp. 1-64.
etectZ'onics~
5.
Kung, S. Y. and Lin, D.W., Optimat Hanket nom model muttivazeiabte systems~ I.E.E.E. Transactions on Automatic Control 26 (1981) pp. 832-852.
zoeductions: 6.
Patel, R.V. and Munro, N., Muttivazeiabte system theozoy and design~ Pergamon Press, Oxford 1981.
7.
Sarason, D., Genezoatized inteppotation in H<»~ Trans. Amer. Math. Soc. 127 (1967) pp. 179-203.
8.
Sarason, D., Operatozo-theozoetic aspects of the NevanlinnaPick inteppolation pzoobtem~ this volume.
9.
Solymosi, J., Inteppo lation l.c7i th PR functions based on method~ Periodica Polytechnica (Budapest) 15 (1971) pp. 71-76.
F. Fenyves' 10.
Yeh, F.B., Nwnezeicat 801.ution of matzei: inteppolation Ph.D. Thesis, Glasgow University 1983.
pzoobtem8~
11.
Young, N.J., The NevanUnna-Pick pzoobtem fozo matzei:z:-vatued to be published.
functions~