PROCEEDINGS
OF
ASIAN L O G I C Conferences
h
This page intentionally left blank
PROCEEDINGS
OF THE
ASIAN L O G I C
h
Conferences 6 - 10 June 1999 29 August - 2 September 2002
Hsi-Tou, Taiwan
Chongqing, China editors
Rod Downey Victoria University of Wellington, New Zealand
Ding Decheng Nanjing University, China
Tung Shih Ping Chung Yuan Christian University, Taiwan
Qiu Yu Hui Southwest China Normal University, China
Mariko Yasugi Kyoto Sangyo University,Japan
associate editor
Guohua Wu Victoria Universiiy of Wellington, N e w Zealand
SINGAPORE UNIVERSITY PRESS NATIONAL UNIVERSITY OF SINGAPORE
Y
World Scientific
N E W JERSEY * L O N D O N * SINGAPORE
SHANGHAI
*
HONG KONG
TAIPEI
BANGALORE
Published by
Singapore University Press Yusof Ishak House, National University of Singapore 31 Lower Kent Ridge Road, Singapore 119078 and World Scientific Publishing Co. Pte. Ltd. 5 Toh Tuck Link, Singapore 596224 USA ofice: Suite 202,1060 Main Street, River Edge, NJ 07661 UK ofice: 57 Shelton Street, Covent Garden, London WC2H 9HE
British Library Cataloguing-in-PublicationData A catalogue record for this book is available from the British Library
PROCEEDINGS OF THE 7TH AND 8TH ASIAN LOGIC CONFERENCES Copyright 0 2003 by Singapore University Press and World Scientific Publishing Co. Pte. Ltd. All rights reserved. This book, or parts thereof, may not be reproduced in any form or by any means, electronic or mechanical, includingphotocopying, recording or any information storage and retrieval system now known or to be invented, without written permission from the Publisher.
For photocopying of material in this volume, please pay a copying fee through the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, USA. In this case permission to photocopy is not required from the publisher.
ISBN 981-238-261-5
Printed in Singapore by World Scientific Printers (S) Pte Ltd
V
PREFACE
The Asian Logic Conference has occurred every three years since its inception in Singapore in 1981. It rotates among countries in the Asia Pacific region with interests in the broad area of logic including theoretical computer science. It is now considered a major conference in this field and is regularly sponsored by the Association for Symbolic Logic. This volume contains papers, many of them surveys by leading experts, of both the 7th meeting in Hsi-Tou, Taiwan, and the 8th in Chongqing, China. A separate volume was in a state of formation after the 7th meeting, when all of the files were lost following a devastating earthquake in Taiwan. Those who had attended the Hsi-Tou meeting were shocked to learn that many of the buildings where we had had our meeting had been completely destroyed. In view of the time that had passed because of consequential delays, a decision was taken to combine the two meetings into one volume. We were very pleased t o find that World Scientific were enthusiastic to support this venture. Authors were invited to submit articles to the present volume, based around talks given a t either meeting. In particular, this allowed for the updating of papers from the 7th meeting. The editors were very concerned to make sure that the planned volume was of high quality. All of the submitted papers were fully refereed, with somewhat over half being accepted for the final volume. We think the resulting volume is fairly representative of the thriving logic groups in the Asia-Pacific region, and also fairly representative of the meetings themselves. For the record, here is a list of main speakers from the two meetings: 8th Asian Logic Meeting: Akihiro Yamamoto (Japan) Sergei Goncharov (Russia) Rod Downey (New Zealand) Bhakhadyr Khoussainov (New Zealand) Robert Goldblatt (New Zealand) Yang Yue (Singapore) Li Angsheng (China) Su Kaile (China) Klaus Weihrauch (Germany) Masahiro Hamano (Japan)
vi
7th Asian Logic Meeting: Felipe Cucker (Hong Kong) Rod Downey (New Zealand) Michael Dunn (USA) Ellery Eells (USA) Byunghan Kim (USA) Ker-I KO (USA) Masanao Ozawa (Japan) Gaisi Takeuti (USA) Stevo Todorcevic (Canada) Akito Tsuboi (Japan) Johan van Benthem (Netherlands) We hope that you enjoy the resulting volume. Sincerely yours, the editors: Rod Downey, Ding Decheng, Tung Shih Ping, Qiu Yu Hui, Mariko Yasugi, and WU Guohua.
vi i
CONTENTS Elementary Properties of Rogers Semilattices of Arithmetical Numberings S. A . Badaev, S. S. Goncharov and A . Sorbi
1
Five Puzzles about Mathematics in Search of Solutions C. S. Chihara
11
Complexity Classes over the Reals: A Logician’s Viewpoint F. Cucker
39
Computability, Definability and Algebraic Structures R. Downey
63
Trivial Reals R. G. Downey, D. R. Hirschfeldt, A . Naes and F. Stephan
103
Popper and Miller, and Induction and Deduction E. Eells
132
Enlargements of Polynomial Coalgebras R. Goldblatt
152
A Layered Approach to Extracting Programs from Proofs with an Application in Graph Theory J. Jeavons, B. Basit, I. Poernomo and J. N. Crossley
193
A Common Structure of Logical and Algebraic Algorithms Y. Kawaguchi
222
Games on Graphs: Automata, Structure, and Complexity B. Khoussainov and T. Kowalski
234
Computational Complexity of Fractals K.-I. KO
252
Vlll
Definability in Local Degree Structures - A Survey of Recent Results Related to Jump Classes
270
A . La and Y. Yang A Limit Stage Construction for Iterating Semiproper Preorders
303
T. Miyamoto An Application of N D J P R O to ~ the Catch and Throw Mechanism M. Nakata, N. Saneto and M. Yasugi The Curry-Howard Isomorphism Adapted for Imperative Program Synthesis and Reasoning I. Poernomo and J. N. Crossley
328
343
Phase-Valued Models of Linear Set Theory M. Shirahata
377
A Problem on Theories with a Finite Number of Countable Models
398
A . Tsuboi Probabilistic Logic Programming with Inheritance J. Wang, S. Ju, X . Luo and J. Hu Sequent Systems for Classical and Intuitionistic Substructural Modal Logics
409
423
0. Watari, T. Ueno, K . Nakatogawa, M. F. Kawaguchi and M. Miyakoshi Diamond Embeddings into the D.C.E. Degrees with 0 and 1 Preserved G. Wu
443
1
ELEMENTARY PROPERTIES OF ROGERS SEMILATTICES OF ARITHMETICAL NUMBERINGS *
S. A. BADAEV Kazakh National University, 39/47 Masanchi Street, Almaty, 480012, Kazakhstan E-mail: badaevOmath.kz
S . S . GONCHAROV Institute of Mathematics of SB R A S , 4 Koptyug Avenue, Novosibirsk, 630090, Russia E-mail: gonchar @math.nsc. ru A. SORB1 Dipartimento di Scienze Matematiche ed Info rmati ch e ”Roberto Mag ari ”, Via del Capitano 15, 531 00 Siena, Italy E-mail:
[email protected]
We investigate differences in the elementary theories of Rogers semilattices of arithmetical numberings, depending on structural invariants of the given families of arithmetical sets. It is shown that at any fixed level of the arithmetical hierarchy there exist infinitely many families with pairwise elementary different Rogers semilattices.
1. Preliminaries and Background
For unexplained terminology and notations concerning computability theory, our main references are the textbooks of A.I. Mal’tsev [l],H. Rogers [2] and R. Soare [3]. For the main concepts and notions of the theory of numberings we refer to the book of Yu.L. Ershov [4]. *This work is supported by grant INTAS-00-499
2
Definition 1.1. A surjective mapping a of the set w of natural numbers onto a nonempty set A is called a numbering of A. Let a and p be numberings of A. We say that a numbering a is reducible to a numbering /3 (in symbols, a p) if there exists a computable function f such that a(n) = P ( f ( n ) )for any n E w . We say that the numberings a and ,B are equivalent (in symbols, a = P ) if a P and /3 a.
<
<
<
S. S. Goncharov and A. Sorbi suggested in [5] a general approach for studying classes of objects which admit constructive descriptions in formal languages. This approach allows us to unify, in a very natural way, various notions of computability and relative computability for different classes of constructive objects. Throughout this paper we will confine ourselves to families of arithmetical subsets of w . We take in this case a Godel numbering {@i}iEw of the first-order arithmetical formulas, and apply this approach as follows, see [5]:
Definition 1.2. A numbering a of a family A of C:+,- sets, with n 2 0, is called C:+,- computable if there exists a computable function f such that, for every m, @ f ( m )is a Cn+l-formula of Peano arithmetic and a(m) = {x E w '32 t= @f(m)(Z)} (where the symbol Z stands for the numeral for 2 and '32 denotes the standard model of Peano arithmetic). The set of C:+,computable numberings of A will be denoted by Com:+,(A).
I
Computable numberings of families of sets which are first-order definable in the standard model of Peano arithmetic are called arithmetical numberings. A family A for which Corn:+, ( A ) # 0 will be called C:+, - computable. If n = 0 then Cy- computable numberings and classical computable numberings of families of c.e. sets coincide. The relation E is an equivalence relation on Com:+,(d) and the reducibility induces a partial order on the equivalence classes of this relation. The equivalence class of a numbering a is called the degree of a , denoted by deg(a). The partially ordered set (Com:+,(A)/=, <) of the degrees of C;+,- computable numberings of A will be denoted by R:+, (A). If a and p are in Com:+,(A), then a new numbering a @ p of A is defined as follows: a @ P(2n) = a(.) and a @ P(2n 1) = ,!?(n),and deg(a @ P ) determines the least upper bound of the pair deg(a), deg(P) in ??,:+,(A). Thus, RO,+,(d)can be regarded as an upper semilattice.
<
+
Definition 1.3. The upper semilattice RO,+,(A)is called the Rogers semilattice of the class of arithmetical numberings of A.
3
The Rogers semilattice RE+,( A )can be viewed as a tool for measuring the algorithmic complexity of computations of the family A as a whole. The theory of computable numberings is mainly with problems related to the algebraic and elementary properties of the Rogers semilattices. We continue the investigation of the elementary types of Rogers semilattices for infinite arithmetical families started in [6],[7] and [8]. We are interested in differences between the elementary theories of Rogers semilattices of families of any fixed level of the arithmetical hierarchy. Everyone who has ever dealt with the classical theory of computable numberings is well aware that general facts about Rogers semilattices of families of c.e. sets are very rare, and at the same time it is very difficult to establish elementary properties that distinguish given structures. In contrast to the classical case, the elementary theories of Rogers semilattices of arithmetical numberings for the level two and higher seem more exciting. In what follows, we briefly examine some algebraic and elementary properties of the Rogers semilattices RE+,(d)for various A. 1.1. Cardinality, Lattice Properties, Undecidability
The following two theorems are well-known facts of the theory of computable numberings in the classical case.
Theorem 1.1. (A. B. Khutoretsky, [9]) For every family A of c.e. sets, if the Rogers semilattice Ry(A) contains at least two elements then it is infinite. Theorem 1.2. (V. L. Selivanov, [lo]) For every family A of c.e. sets, if the Rogers semilattice Ry(A) contains at least two elements then it is not a lattice. For Cy-families, Theorems 1.1and 1.2 answer questions posed by Yu. computable families have been answered by Goncharov and Sorbi, [5].
L. Ershov. The corresponding questions for the case of C;+,-
Theorem 1.3. If a CE+2-computable family A contains at least two (A) is infinite and is not a lattice. elements then the Rogers semilattice Theorem 1.3 gives us a complete description of the families A C CE+, since if A consists of a single element then all numberings of A are evidently equivalent. For details relative to the classical case, we refer to [ll],and we recall the following well-known problem raised by Yu. L. Ershov.
4
Question 1.1. Under what conditions is the Rogers semilattice R!(A)of a family of c.e. sets non-trivial?
It should be noted that the elementary theory of R:+, (A)of every nontrivial family A is quite complicated. We give evidence for this statement as follows. Let E* denote the bounded distributive lattice obtained by dividing the lattice E of all c.e. subsets of w modulo the ideal of all finite sets. We will denote by fi the principal ideal of R:+,(A),
P + {dedy) I deg(7) ,< deg(P)l. Theorem 1.4. (S. Yu. Podzorov, [12], see also [7]) Let A be any X:+,computable family. There exists a numbering a E Com:+,(A) such that (1) ii is isomorphic to ~*\{l} if the family A is infinite;
(2) ii is isomorphic to E* if the family A is finite. Theorem 1.4 and the fact that the elementary theory of E* is hereditarily undecidable, [13], immediately yield: Corollary 1.1. The elementary theory of every non-trivial Rogers semilattice R:+,(A) is hereditarily undecidable. Theorem 1.4 and Corollary 1.1give us a deep insight into the complexity of Rogers semilattices of Xi+,-computable families. The case of Eycomputable families is still open: Question 1.2. Is the elementary theory of any non-trivial Rogers semilattice of a Ey - computable family hereditarily undecidable, or at least undecidable? 1.2. Extremal Elements What kind of computable numberings should be thought of as the most natural ones? A partial answer to this question, as well as a motivation for introducing the notion of a universal numbering, is given by the next proposition [6]. Proposition 1.1. Let a be a numbering of a family A 5 X:+,. Then the following statements are equivalent: (i) a is X:+, - computable; (ii) a is reducible to the numbering WO'"' of the family of all C:+,-sets; (iii) a is O ( n ) - reducible to WO'"' .
5
Definition 1.4. A numbering (I! of A C C:+, is called principal or universal in Corn:+,(A) if (i> E corn:+, ( 4 , (ii) /3 6 (I! for all numberings /3 E Corn:,, (A).
It is obvious that the greatest element, if any, of the Rogers semilattice of any family A is exactly the degree of some universal numbering of A. Proposition 1.1implies also that many essential facts and notions relative to universal numberings are easily lifted from principal computable numberings of families of c.e. sets to arithmetical numberings. For instance, 0 0
0
Ershov's classification of principal subsets, see [4]; the closure condition of Lachlan [14] for families of sets to have computable principal numberings; existence of universal numberings in Corn:+, (A) with respect to O(n)- reducibility, for every finite family A E C:+, .
In particular, we should mention the following two examples which show a difference between the Rogers semilattices of some infinite families.
Example 1.1. The family C:+, of all C:+,-subsets of w has a universal numbering in Corn:+, namely the relativization WO'"' of the classical Post numbering W of the family of all c.e. sets. Example 1.2. For every n, the set F of all finite sets is obviously C:+,computable and has no universal numbering in Com:+,(F). The latter holds by the relativized version of Lachlan's condition, [14]: if any C:+,computable family has a universal numbering then it is closed under unions of increasing C:+,- computable sequences of its members.
These examples show an elementary difference between RE+,(C:+,) and R:+,(F)with regard to the existence/non-existence of the greatest element in these semilattices. As regards finite families, examples of elementary differences between Rogers semilattices of finite families are provided by the following result of S. A. Badaev, S. S. Goncharov, and A. Sorbi, [6],
Theorem 1.5. Let A Xi+, be a finite family. Then A has an universal numbering in Com:+,(A) if and only if A contains a least element under inclusion.
6
Again, as in the examples above, existence/non-existence of the greatest element provides an elementary property which allows us to distinguish some Rogers semilattices. To compare elementary properties of Rogers semilattices of finite families versus Rogers semilattices of infinite families, we can use a different type of extremal element, namely minimal elements of the semilattices. It is a well-known fact in the theory of numberings that any finite family has a numbering which is reducible to all the numberings of that family, see [4].This fact does not depend on either the nature of the family or the computability of the considered numberings. Thus, the Rogers semilattice RO,+,(d) of any finite family d of C:+,-sets has a least element. On the other hand, we have the following theorem of S. A. Badaev and S. S. Goncharov, [151.
Theorem 1.6. For every n, i f A is an infinite C;+,then R:+2 (A) has infinitely m a n y minimal elements.
computable family,
Remark 1.1. Theorem 1.6 does not hold for some infinite families of c.e. sets and does hold for other infinite families of c.e. set. Furthermore, the following question is a problem of Yu. L. Ershov known since the 60’s. We refer to [ll]for details of this problem. Question 1.3. What is the possible number of minimal elements in the Rogers semilattice Ry(d) of a family of c.e. sets? 1.3. The Weak Distributivity Property
In this subsection we are concerned with an interesting and natural elementary property of Rogers semilattices which establishes one more difference between RO,+,(d),with d finite, and RO,+,(B), with B infinite. We refer to [8] for details and proofs. First we recall some definitions.
Definition 1.5. An upper semilattice ( L ,V, <) is called distributive if for every a l , a2, b E L , if b 5 a1 V a2 then there exist b l , b2 6 L such that bl 5 a l , b2 5 a2 and b = bl V b2. Theorem 1.7. For every n and for every finite family A & C:+, , Rz+,(A) is a distributive upper semilattice. The situation is different if we consider infinite families. First of all, we notice:
7
Remark 1.2. It is easy to see that the three element upper semilattice LO = { a , b, c}, where a and b are incomparable and c = a V b, is not distributive. There exist many Rogers semilattices which contain LOas an ideal, [4],and, therefore, are not distributive. However, if we add I to LO, we do obtain a distributive lattice. This remark motivates our next definition.
Definition 1.6. An upper semilattice C = ( L ,<) is weakly distributive if 21 = ( LU {I},
1
\
L u {I}}.
Proposition 1.2. An upper semilattice ( L ,V, <) is weakly distributive if and only if f o r every a l , a2, b 6 L, if b 5 a1 V a2 and b $ a l , b $ a2 then there exist bl ,b2 E L such that bl 5 a l , bz 5 a2 and b = bl V b z . Theorem 1.8. For every n, the Rogers semilattice of any infinite C0,+,computable family is not weakly distributive. Question 1.4. Does there exist a computable infinite family A of c.e. sets such that R:(A) is distributive? Does there exist a computable infinite family A of c.e. sets such that Ry(d)is weakly distributive? 2. The Main Result
It should be noted that Rogers semilattices of families from different levels of the arithmetical hierarchy can be surprisingly different, as can be seen from the following theorem of S. A. Badaev, S. S. Goncharov and A. Sorbi,
PI. Theorem 2.1. For every n there exists m 2 n and a C;+,- computable family B such that no Rogers semilattice RO,+,(d)of any C;,, - computable family A is isomorphic to RO,+,(B). The differences between Rogers semilattices established in Theorem 2.1 are based on the fact that ideals of Rogers semilattices of families chosen from different levels of the arithmetical hierarchy have different algorithmic complexities. Unfortunately, these differences are not elementary. So Theorem 2.1 provides a natural motivation for searching elementary properties between Rogers semilattices of families lying in the same level of the arithmetical hierarchy. Some fruitful ideas from the paper of V. V. V’jugin, [16] were very useful for our research.
8
Theorem 2.2. For every k E w, there exist infinitely many Xi+,-computable families with elementay pairwise different Rogers semilattices.
Sketch of proof. Let k be an arbitrary natural number. We will construct a sequence {Be},>l - of infinite Ci+,-computable families such that ~ h ( ~ i(Bet)) + i
# ~ h ( ~ i( B+e l il ) )
for all e’ # e”. Indeed we will construct a sequence {A,},?l of families of sets of which {Be}e21 is subsequence. Let M stands for any O(’)- maximal set, and let n be a natural number. Let EA, E:, . . . ,E,” be a computable partition of w into infinite computable sets. Let f : denote some arbitrary computable bijection of w onto EA,i E [ l , n ] .Clearly, Mi + Z i U f:(M) is a O(k)-maximal set for each i E [l,n]. i E [ l , n ] stand , for the family { M i U { z } 1 z E The Let families A; are evidently C!+,-computable. Define A, i= U A;.
ai}.
i~[l,n]
Lemma 2.1. For all numbers n > 0 and i E [ l , n ] ,every numbering a E Comi+,(A,), and every set A E A;, the index sets with respect to a of the subfamilies A; and { A } are C;,, - sets. Lemma 2.2. For all numbers n > 1, i E [1,n] and all numberings v,vo,v1 E Com:+,(d,), if Y = YO @ v1 then all but finitely many sets of A: are contained either in V O ( W ) or in v~(w). Lemma 2.3. For all numbers n > 0 and m 2 n 0 7:,7LY;7,-r;7...,7:+1,7k+l Com,+,(A,), i f 7 b ,-rm+l 0 @ 7;+, then there ex& a numbering 6 E Corn:+, sequence ~1~~ 2 . ., . ,cm such that 6 7:+, and S 7:’
<
<
and all numberings : =7;@7; = * * . (A,) and a banary @ 7;’ CB. . . @ 7Lm.
Definition 2.1. We will say that any two Xi+,-computable numberings YO,y of a family A induce a minimal pair in the Rogers semilattice RE+,(A) if there is no numbering v E Com;+,(A) such that v vo and v v1.
<
<
In the proof of Lemma 2.3 we construct two numberings which do not induce a minimal pair in Ri+l(An ). On the other hand we consider.now some regular way of constructing numberings which induce minimal pairs in Rogers semilattice Ri+,(An).
9
For every i E [l,n], we fix two different numbers a!, at E numberings a:, s 5 1 as follows: for every x , we let
a:(.)
*
{
Mi and define
Mi U{aq}, if x E M, Mi U{f h ( x ) }otherwise.
It is obvious that a: E Com:+,(di). Lemma 2.4. RE+, (41.
The numberings a: and at induce a minimal pair in
Lemma 2.5. For every m > 0 and n 2 22m+’, there exist numberings PY,P;,Pg,Pl,.-. ,P,”m,PL E Comi+l(An) such that c%& &a Pi z.. . 0,”m Pim; for every i E [1,2m], the numberings and P: induce a minimal pair in Xi+, (An); for every 1 5 m, every set I = {il < ig < . . . < il} [l,2m], every binary sequence u1, ug, . . . , ( T I , and every E E {0,1} and i E [I,2”] \ I, the numberings PI and P z @ P@ : , . . . ,@Pz induce a minimal pair in RiCl(An). Using Lemmas 2.1-2.5 we can now deduce the statement of the theorem
+
as follows. Define a computable function h by letting h(1) = 16 and h(e h(e)+1 1) = 22 for every e 2 1. Let Be + for every e 2 1. Lemmas 2.3, 2.5 imply that T h ( R ~ + , ( B , ~#)T ) h ( R ~ + l ( B efor ~ ~every ) ) e‘ # e”. References
1. A. I. Mal’tsev, Algorithms and Recursive Functions. Nauka, Moscow, 1965 (Russian); Wolters-Noordoff Publishing, Groningen, 1970 (English translation). 2. H. Rogers, Jr., Theory of Recursive Functions and Eflectiue Computability. McGraw-Hill, New York, 1967. 3. R. I. Soare, Recursively Enumerable Sets and Degrees. Springer-Verlag, Berlin Heidelberg, 1987. 4. Yu. L. Ershov, Theory of Numberings. Nauka, Moscow, 1977 (Russian). 5. S. S. Goncharov and A. Sorbi, Algebra and Logic, 36,359-369 (1997). 6. S. A. Badaev, S. S. Goncharov and A. Sorbi, In: Computability and Models. Kluwer Academic/Plenum Publishers, Dortrecht, 11-44, 2002. 7. S. A. Badaev, S. S. Goncharov, S. Yu. Podzorov and A. Sorbi, In: Computability and Models. Kluwer Academic/Plenum Publishers, Dortrecht, 4577, 2002.
10
8. S. A. Badaev, S. S. Goncharov and A. Sorbi, In: Computability and Models. Kluwer Academic/Plenum Publishers, Dortrecht, 79-91, 2002. 9. A. B. Khutoretsky, Algebra and Logic, 10,348-352 (1971). 10. V. L. Selivanov, Algebra and Logic, 15,297-306 (1976). 11. S. A. Badaev and S. S. Gonchaxov, Contemporary Mathematics, 257, 23-38 (2000). 12. S. Yu. Podzorov, Algebra and Logic, to appear. 13. L. Harrington, A. Nies, Adu. Math., to appear. 14. A. H. Lachlan, Zeit. Mat. Log. Grund. Math., 10, 23-42 (1964). 15. S. A. Badaev and S. S. Goncharov, Algebra and Logic, 40, 283-291 (2001). 16. V. V. V’jugin, Algebra and Logic, 12, 277-286 (1973).
11
FIVE PUZZLES ABOUT MATHEMATICS IN SEARCH OF SOLUTIONS
CHARLES S. CHIHARA Department of Philosophy University of California Berkeley, CA 94720 U.S. A emad: charlesl@socrates. Berkeley.EDU
This paper will be delivered in two parts. In the first, I shall describe five puzzles about mathematics that cry out for solutions. In the second, I shall sketch a view of mathematics in terms of which analyses and explanations of these puzzles will be given. That solutions to these puzzles can be found from the perspective of this view of mathematics provides some support for the view itself.
I. The Five Puzzles [l]A puzzle about geometry
Consider the first three postulates of Euclid’s version of plane geometry: Postulate 1: A straight line can be constructed from any point t o any point. Postulate 2: A straight line can be extended indefinitely in a straight line. Postulate 3: A circle can be constructed with its center at any point and with any radius. a aThe first three postulates of Euclid’s geometry are translated by Thomas Heath (in (Heath, 1956), p. 154) as follows: Let the following be postulated: 1. To draw a straight line from any point to any point. 2. To produce a finite straight line continuously in a straight line. 3. To describe a circle with any centre and distance.
But he understands Postulate 1 to be “asserting the possibility of drawing a straight line from one point to another”, Postulate 2 to be “maintaining the possibility of producing a finite straight line continuously in a straight line”(p. 196), and in the case of Postulate 3,
12
Now compare those postulates with the first three axioms of Hilbert’s version published in his Foundations of Geometryb. Axiom 1: For every two points A , B there exists a line L that contains each of the two point A, B. Axiom 2: For every two points, A, B there exists no more than one line that contains each of the points A, B. Axiom 3: There exists at least two points on a line. There exists at least three points that do not lie on a line. Notice that Hilbert’s axioms are existential in character: they assert the existence of certain geometric objects, i.e. points and lines. Euclid’s postulates, on the other hand, do not assert the existence of anything. Rather, what is asserted is the possibility of making some sort of geometric construction‘. In contrast to Hilbert’s existential geometry, then, Euclid’s he tells us that “Euclid’s text has the passive of the verb: ‘a circle can be drawn’. . .”(p. 199). Commenting on Postulate 4, Heath notes that, according to Proclus, “Geminus held that this Postulate should not be classed as a postulate but as an axiom, since it does not, like the first three Postulates, assert the possibility of some construction but expresses an essential property of right angles”(p. 200). E. G. Kogbetliantz (in (Kogbetliantz, 1969), p. 554) gives the straightforwardly modal translation of Euclid’s postulates as follows: Postulate 1: A straight line can be drawn from any point to any point. Postulate 2: A straight line may be produced, that is, extended indefinitely in a straight line. Postulate 3: A circle can be drawn with its center at any point and with any radius. Peter Wolff explains Euclid’s first three postulates as follows: “The root meaning of the word ‘postulate’ is to ‘demand’; in fact, Euclid demands of us that we agree that the following things can be done: that any two points can be joined by a straight line; that any straight line may be extended in either direction indefinitely; that given a point and a distance, a circle can be drawn with that point as center and that distance as radius”((Wolff, 1963), pp.47-8). b(HiIbert, 1971),pp.3-4. ‘Cf. Mueller’s comments (in, p. 14): “Hilbert asserts the existence of a straight line for any two points, as part of the characterization of the system of points and straight lines he is treating. Euclid demands the possibility of drawing the straight-line segment connecting the two points when the points are given. This difference is essential. For Hilbert geometric axioms characterize an existent system of points, straight lines, etc. At no time in the Grundlagen is an object brought into existence, constructed. Rather its existence is inferred from the axioms. In general Euclid produces, or imagines produced, the objects he needs for a proof, . . . It seems fair to say that in the geometry of the Elements there is no underlying system of points, straight lines, etc. which Euclid attempts to characterize.” My colleague Paolo Mancosu has informed me that his researches indicate that some Greek mathematicians read the modal Euclidean postulates as having an existential commitment to eternal objects and that, even at the time of Euclid, there were disputes about the “ontological commitments” of the postulates of geometry. It would seem that the historical facts are more complicated than I have indicated above, but my basic point is that there is a way of understanding the postulates of Euclid’s
13
geometry is modal in character. For over two thousand years, geometry was understood and developed by many mathematicians as a modal theory, but for some reason, at some time before the 20th Century, geometry became straightforwardly existential. Hilbert was by no means the first mathematician to think and reason about geometrical objects (such as points and lines) in terms of existence rather than constructibility. Mathematicians had begun to shift to the existential mode of expressing geometrical theorems hundreds of years before Hilbert had written on the topic. This shift in geometry from making constructibility assertions to asserting existence raises some fundamental questions:
(a) No one seems t o have made a fuss about the change that took place or to have even taken note of it. No one believes that an ordinary existential statement such as ”There are buildings with over three hundred stories” is equivalent t o a modal statements of the form “It is possible t o construct buildings with over three hundred stories”. Why weren’t there serious debates among mathematicians and philosophers over the validity of making such an apparently radical ontological change in the primitives of one of the central theories of mathematics? (b) The applications t o which geometry was put evidently did not change as a result of the described shift. How is it that, in our reasoning about areas, volumes, distances, etc., it seems t o make no difference whether the geometry we use asserts what it is possible t o construct or whether it asserts the existence of mathematical entities?
A completely adequate answer to these questions would require a detailed investigation into the history of mathematics spanning many hundreds of years-something that is well beyond the scope of this paper. So, I shall concentrate my investigation to the following (restricted but closely related) puzzle about Hilbert’s view of geometry:
geometry that is such that no commitment to mathematical objects is presupposed by them and that some mathematicians and philosophers understood them in that way.
14
In the introduction to his Foundations of Geometry, Hilbert writes: “The establishment of the axioms of geometry and the investigation of their relationships is a problem . . . equivalent to the logical analysis of our perception of space”(Hilbert, 1971 #287, p. 2). And he goes on to say that his axioms express “facts basic to our intuition”(p. 3). But does our perception of space tell us that there exist the infinity of points and lines postulated by his geometrical axioms? Does our intuition inform us that such points and lines truly exist? How could Hilbert have felt justified in postulating axioms that express such existential “facts”, when Euclidean plane geometry had been developed and applied for a multitude of centuries without any such commitment to an apparent ontology of imperceptible objects?
I now turn to the second of my puzzles. [2] The inertness of mathematical objects
As Jody Azzouni has emphasized in his book (Azzouni, 1994), mathematical practice supports a view according to which mathematical objects are very different in kind from the objects the empirical scientists study: A crucial part of the practice of empirical science is constructing means of access to (many of) the objects that constitute the subject matter of that science. Certainly this is true of theoretical objects such as subatomic particles, black holes, genes, and so on. Empirical scientists attempt to interact with most of the theoretical objects they deal with, and it is almost never a trivial matter to do so. Scientific theory and engineering know-how are invariably engaged in such attempts, which are often ambitious and expensive. Nothing like this seems to be involved in mathematics. ((Azzouni, 1994), p.5.)
Continuing this line of thought, it should be noted that mathematicians do not give seminars or conference presentations on the newest means of detecting sets, numbers, or functions. Detecting or interacting with mathematical objects does not seem to constitute any part of our mathematical practice. It would seem that the objects of mathematics do not interact with us or with anything in our world. This inertness feature of mathematical objects, indicated by the contrast between our mathematical and scientific practices, gives rise to a pair of closely related philosophical questions: (1) If mathematical objects are not detectable by our best scientific instruments, how are we able to refer to them? (2) How are we able to gain knowledge of these inert mathematical
15
objects?d These naturally arising questions have engendered a considerable literature. Since time is short and there is much ground to be covered in this paper, I shall not attempt to describe the various attempts that have been made to answer the above two philosophical questions. I shall simply give you my considered judgment that none of the answers that have been proposed thus far have even remotely satisfied me. So from my point of view, we have a fundamental puzzle about the nature of mathematics, in need of analysis and explanation. Let us consider now the third of my puzzles.
[3] Consistency and mathematical existence
If one looks back at the history of mathematics, one finds a string of brilliant mathematicians making what seems to be the same basic point, viz. that mathematical existence is a matter of consistency. Consider what Henri Poincare wrote in Science and Hypothesis: The word ‘existence’ has not the same meaning when it refers to a mathematical entity as when it refers to a material object. A mathematical entity exists provided there is no contradiction implied in its definition, either in itself, or with the proposition previously admitted. ((Poincarb, 1952), p. 44).
The above point is amplified in Science and Method, where he wrote: If
.
. . we have a system of postulates, and if we can demonstrate that
these postulates involve no contradiction, we shall have the right to consider them as representing the definition of one of the notions found among them. ((PoincarC, 1953), p. 152).
The following quotation from the writings of David Hilbert expresses a thought that is close t o what Poincar6 expressed in the previous quotations:
dSee (Azzouni, 1994), p.5.
16
If contradictory attributes are assigned to a concept, I say, that mathematically the concept does not exist. So, for example, a real number whose square is -1 does not exist mathematically. But if it can be proved that the attributes assigned to the concept can never lead to a contradiction by the application of a finite number of logical inferences, I say that the mathematical existence of the concept (for example, of a number or a function which satisfies certain conditions) is thereby proved. In the case before us, where we are concerned with the axioms of real numbers in arithmetic, the proof of the consistency of the axioms is at the same time the proof of the mathematical existence of the complete system of real numbers or of the continuum.
A third outstanding mathematician seems to have maintained a view that is very similar to the above. According to the historian of mathematics Joseph Dauben, Logical consistency was the touchstone that Cantor applied to any new theory before declaring it existent and a legitimate part of mathematics . . .. Since he took his transfinite numbers to be consistently defined, . . ., there were no grounds to deny his new theory. This kind of formalism, stressing the internal conceptual consistency of his new numbers, was all mathematicians needed to consider before accepting the validity of the transfinite numbers. ((Dauben, 1990), p. 129).
The above three are by no means the only mathematicians who have expressed such thoughts. Indeed, such an idea is considered a commonplace by some. For example, Paul Bernays has written: “It is a familiar thesis in the philosophy of mathematics that existence, in the mathematical sense, means nothing but consistency.”f For my purposes, these examples should suffice. Now it would be surprising, wouldn’t it, if there were absolutely nothing in what all these outstanding mathematicians were claiming. After all, these researchers were both brilliant and knowledgeable about mathematics, and it would be implausible to attribute their views about mathematical existence to some extreme philosophical position, since they held widely differing philosophical views about the nature of mathematics.g eThis quotation is from an excerpt of Hilbert’s famous address to the International Congress of Mathematics of 1900, in which he put forward his list of 23 important unsolved problems, published in (Calinger, 1982), p. 662. ‘From a translation of (Bernays, 1950) by Richard Zach. BPoincard was a predicativist, Hilbert founded Formalism, and Cantor was a Mathematical Platonist.
17
The puzzle is: W h a t m u s t mathematics be like if the mere consistency and coherence of the definition of a totality of mathematical objects seems t o be suficient to a number of brilliant practitioners of the science t o yield the acceptability of the reality of such objects? This takes me to the fourth of my puzzles: [4] Different attitudes of practicing mathematicians regarding the ontology of mathematics Some outstanding mathematicians have believed that the set theorist is reasoning about some non-physical entities that truly exist (I have in mind such Platonic or Realist researchers in set theory as Kurt Godel, Robert Solovay, and John Steel), whereas other outstanding mathematicians, such a Alfred Tarski, Paul Cohen, and Abraham Robinson, have maintained that set theorists are not reasoning about things that truly exist at all, with Cohen opining that “probably most of the famous mathematicians who have expressed themselves on the question have in one form or another rejected the Realist p ~ s i t i o n ” ~What . is striking about the latter group is that, despite such skeptical beliefs about the ontology of mathematics, many of them have produced, and continue to produce, important and fruitful mathematical results. This is clearly very different from what is typical in the other sciences. A chemist, who does not believe in phlogiston, does not theorize in the phlogiston theory, perform experiments based upon the theory, and explain phenomena with the theory. Few, if any, chemists who were skeptical about the existence of phlogiston continued to produce fruitful developments in phlogiston theory.’ Similarly, there do not exist large groups of outstanding geneticists, who deny that there are such things as genes. How is it, then, that there are so many mathematicians who are thoroughly skeptical of the existence of mathematical objects, and yet continue to work fruitfully in such fields as set theory? Evidently, there is something about the nature h(Cohen, 1971), p. 13. Robinson has expressed his anti-realist views in (Robinson, 1965); Cohen in (Cohen, 1971). I personally heard Tarski express his nominalistic views on several occasions in Berkeley. Other outstanding mathematicians at Berkeley who reject the Realist position are Jack Silver and John Addison. ‘There may be some areas of physics, such as quantum mechanics, in which significant numbers of researchers are skeptical about the existence of the theoretical entities that the principal theories of the area postulate. But such areas, I contend, would not be typical.
18
of mathematics that fosters such behavior. So the puzzle is: how does the science of mathematics differ from the other sciences that accounts for this striking difference? The last puzzle I take up is one that I owe to metaphysicians. [5] The van Inwagen Puzzle
This is a puzzle about set theory put forward by the philosopher Peter van Inwagenj We are to imagine that a philosopher has advanced a theory of typosynthesis that makes the following assertions: (1) There are exactly ten cherubim. (2) Each human being bears a certain relation, typosynthesis, to some but not all cherubim. (3) The only things in the domain of this relation are human beings. (4) The only things in the converse domain of this relation are cherubim.
This is all the typosynthesis theory says.k The theory is concerned solely with cherubim and the relation of typosynthesis that humans bear to cherubim and all it says about the cherubim is given by the four assertions above. To understand van Inwagen’s puzzle, one needs to grasp some of the basic concepts he used to pose his puzzle. We need, first, to distinguish intrinsic from extrinsic relations. We must also distinguish two kinds of intrinsic relations: internal and external relations. To characterize these intrinsic relations, we need to understand what an intrinsic property is. The intrinsic properties of a thing are those properties the thing has just in virtue of the way it is - not in virtue of any relation it is in with other things. For example, you have a heart, a kidney, you are covered with skin, you have hair, etc. Thus, you can be said to have the intrinsic properties of having a heart, having a kidney, of being covered with skin, of having hair, etc. The property of being married, on the other hand, is not one of your intrinsic properties, since you have that property in virtue of being in a certain relation to another person (your spouse). The property of being married is classified by philosophers as one of your extrinsic properties present van Inwagen’s ideas on this topic as a puzzle, but in fact his paper was intended to be a defense of Alvin Plantinga’s Modal Realism against an objection raised by David Lewis. For a discussion of Lewis’s objection and van Inwagen’s defense, see (Chihara, 1998), Chapter 3, Section 9. k(van Inwagen, 1986), p.206.
jI
19
properties that a thing has in virtue of its relation or lack of relation to other things. Now how do we distinguish the two kinds of intrinsic relations: the internal and external relations? Internal relations are relations that “SUpervene” on the intrinsic properties of the relata. What does this mean? To assert supervenience is to deny independent variation. As David Lewis describes it: “To say that so-and-so supervenes on such-and-such is to say that there can be no difference in respect of so-and-so without difference in respect of such-and-such” ((Lewis, 1983), p. 358). Thus, what we can infer from these definitions is this: if X - 1 and Y - 1 are in the internal relation R, but X - 2 and Y - 2 are not, then necessarily, either the X’s have different intrinsic properties or the Y’s do (or both). Thus, suppose that there were possible people who were exact duplicates of you and your wife (insofar as one had all the intrinsic properties that you have and the other had all the intrinsic properties that your wife has).’ Then (assuming that you are, in fact, taller than your wife), since necessarily your duplicate would be taller than the duplicate of your wife, we can infer that the relation of being taller than is an internal relation. On the other hand, a relation is external if its holding depends on the intrinsic properties of the composite of the relata, but not just of the relata themselves. To use an example Lewis gives, the relationships of distance holding between the electron orbiting and the proton of a classical hydrogen atom are not internal, since these relations do not just depend upon the intrinsic properties of the electron and the proton taken separately. But if we take the composite - the hydrogen atom - then the relations holding does depend upon the intrinsic properties of the composite. A relation that is not intrinsic is said to be extrinsic. An extrinsic relation, then, is one that is neither internal nor external. An example of an extrinsic relation is that of being an owner of a person is related to a piece of property by this relation. Now consider again the typosynthesis theory. It says that each human being is related by typosynthesis to some cherubim. IS this relation of typosynthesis an intrinsic relation? If so, it must be either internal or external. Suppose the former. Then there must be something about the ‘One might wonder how anything can have a duplicate, since it would be reasonable to suppose that one intrinsic property any thing x would have is the property of being identical to x. So it would be more accurate to define a duplicate in terms of intrinsic qualitative properties -where a qualitative property is what Adams calls a “suchness” (as opposed to a “thisness”). See, in this regard, (Adams, 1979)
20
intrinsic properties of the human and the cherubim in virtue of which the relation obtains. But we have no idea of what intrinsic properties a cherub has. Do they have wings? Are they in physical space and time? Do they have thoughts? Are they changeable? The theory doesn’t say. Clearly, we are in no position to say what intrinsic properties a cherub has in virtue of which a particular cherub is related by typosynthesis to a particular human being. So if the relation were internal, we wouldn’t have the vaguest idea of which internal relation it was. Perhaps we should classify typosynthesis as an external relation. In that case, we can bring into consideration the intrinsic properties of the composite of a human and a cherub. But what is it about the intrinsic properties of the composite in virtue of which the relation holds? Again, we have no idea at all. Adding a composite to the situation does not help. Can we classify the relation of typosynthesis as extrinsic? In that case, there would have to be some other objects in the universe in virtue of which some particular human was in the typosynthesis relation to some cherub. Do we have any idea of what this thing could be? Not at all. It seems then that we have no clear idea of what this relation of typosynthesis is. In that case, says van Inwagen, it must be by magic that we understand what ‘typosynthesis’ means. Such a philosophical theory must be rejected as philosophically unsatisfactory.
mSome philosophers have objected to the above reasoning on the grounds that the distinctions being used are vague and “metaphysical”. Still, the main idea of the objection to the typosynthesis theory can be discerned, even in the absence of these distinctions. Consider some typical relations, say the relations taller than and weighs more than. We know, in general, what features or properties of John and Mary must be taken into account to determine if John is taller than Mary or whether John weighs more than Mary. But what properties of Hillary Clinton and some Cherub must be taken into account to determine if Hillary Clinton is in the relation typosynthesis to the Cherub? Who knows? The above theory does not tell us anything about this relation other than some things about humans being in the relation to some Cherubs. In virtue of what properties of Hillary Clinton and what features of some Cherub is Hillary related by typosynthesis to that Cherub? The answer is: We haven’t the slightest idea. Perhaps, typosynthesis is not that sort of relation. Perhaps, typosynthesis is like the relation of being married to. Perhaps it is in virtue of something Hillary Clinton, the Cherub, or some third being has done that brings it about that they are in the relation in question. Again, we haven’t the vaguest notion of what actions, if any, are required for the relation to obtain. Perhaps typosynthesis holds of Hillary Clinton and the Cherub in virtue of the relative spatial or temporal relationships they have to one another. But if so, we have no idea of what these relationships could be. Thus, it is hard to see how we can have anything like a true understanding of this relation of typosynthesis. And so it would seem that the theory of typosynthesis can hardly be a satisfactory one.
21
Why bring this example in? Because essentially the same things can be said about set theory. As in the typosynthesis case, the set theorist cannot tell us what sort of relationship membership is. Thus, consider the enormous totality of unit sets that are supposed to exist. Only one of these unit sets is the one that has, as its only member, Bill Clinton. Then, what properties of Bill Clinton and this singleton determine that it is Bill Clinton and nothing else that is in the membership relation to this unit set? Who knows? Set theory does not tell us. Perhaps, membership is not that sort of relation. Perhaps, membership is like the relation of being married to. Perhaps it is in virtue of something Bill Clinton, the singleton, or some third being has done that brings it about that Bill is in the membership relation to the set in question. Again, we haven’t the vaguest notion of what actions, if any, are required for the relation to obtain. Perhaps typosynthesis holds of Bill Clinton and the singleton in virtue of the relative spatial or temporal relationships they have to one another. But if so, we have no idea of what these relationships could be. It is hard to see how we can have anything like a true understanding of this relation of membership. And so it would seem that set theory can hardly be a satisfactory one. For those who have some understanding of the metaphysical distinctions with which van Inwagen reasons, one can argue that, as in the typosynthesis case, the set theorist has such a poor grasp of the membership relation that she cannot classify the relationship that an object has to its singleton as intrinsic or extrinsic, internal or external. Thus, consider again the question as to whether the relation that Clinton bears to its singleton an internal relation? If so, there must be something about the intrinsic properties of Clinton and his singleton in virtue of which the membership relation holds. Well what is it about the intrinsic properties of just that one unit set in virtue of which Clinton is related to just it and not to any of the other unit sets in the set theoretical universe? We haven’t the vaguest idea. We know no more about the intrinsic properties of sets than we know about the intrinsic properties of cherubim. Set theory gives us no information about the intrinsic properties of sets, and we are not in any sort of causal relationship with sets, whereby we could learn something about the intrinsic properties of sets by empirical means. So it is a complete mystery how any one could have understood this relation. The conclusion seems to be that set theory is an unsatisfactory theory that should be rejected, just as we concluded above for the case of the typosynthesis theory. But van Inwagen was reluctant to draw that conclusion. He could not bring himself to conclude that set theory is an unsatisfactory
22
theory because he felt that would imply that mathematics is an unsatisfactory theory that should be rejected - for him, a complete absurdity. So he concluded instead that there must be something wrong with his reasoning. David Lewis, the Princeton philosopher upon whose reasoning van Inwagen modeled his typosynthesis reasoning, agrees that the above considerations do seem to lead to the conclusion that we do not understand the primitive membership relation of set theory, writing: It’s a nasty predicament to claim that you somehow understand a primitive notion, although you have no idea how you could possibly understand it. That’s the predicament I’m in . . . . ((Lewis, 1991), p. 36).
But Lewis goes on to say (in apparent agreement with van Inwagen’s sentiments) that he cannot accept the conclusion that he does not grasp the membership relation of set theory, for that implies, he believes, that he should reject “present-day set-theoretical mathematics” ((Lewis, 1991), p. 36). He concludes: If there are no classes, then our mathematical textbooks are works of fiction, full of false ‘theorems’. Renouncing classes means rejecting mathematics. That will not do. ((Lewis, 1991), p. 58).
“Hot though it is in the frying pan, that fire is worse” he says ((Lewis, 1991), p. 36). He tells us that he is moved to laughter “at the thought of how presumptuous it would be to reject mathematics for philosophical reasons”((Lewis, 1991), p. 59). “How would you like the job of telling the mathematicians” he continues, “that they must change their ways, and abjure countless errors, now that philosophy has discovered that there are no classes?” “Not me” says Lewis: so he continues to maintain his belief that he somehow grasps the fundamental primitive of set theory, even though he has no idea how he could understand it. In summary, van Inwagen concludes that there must be something wrong with the argument, but he doesn’t know what, whereas Lewis concludes that he must grasp the membership relation, but he doesn’t know how. In either case, we are left with a real puzzle.
11. Analysis and Explanation
I shall now offer my analyses and explanations of these five puzzles, taking them up in reverse order, starting with:
23
The van Inwagen puzzle
I shall begin my discussion of this puzzle by investigating van Inwagen’s reasoning in support of his belief that, since we cannot classify the membership relation as internal, external or extrinsic, we can apparently infer that we do not grasp the membership relation and hence do not understand mathematics. Of course, van Inwagen could not accept such an inference, and concluded that there must be something wrong with his reasoning, even if he knew not what. But should we be strongly tempted to draw such an inference? Let us ask: What is it to understand mathematics? Here, we should restrict our considerations to just some area of mathematics, say set theory. B y the usual criteria that we use in American universities to determine i f someone understands a substantial amount of set theory, we see if the person can explain the principal concepts of the theory (by giving the relevant definitions, applying these definitions within the theory, and explaining their implications f o r the theory), knows the axioms or fundamental assumptions of the area, and also can cite, prove, explain and applp ( i n a variety of set theoretical contexts) the principal theorems of the area (both basic and advanced). Nothing in the van Inwagen argument shows that mathematicians do not understand mathematics according to these criteria. Thus, I agree with van Inwagen’s reasoning to this extent: if his reasoning truly showed that no one understands mathematics according to the above criteria, then we really would have an absurdity. But it doesn’t. So what is going on? Let us reexamine the typosynthesis theory. Suppose that we regard the theory not as a theory about particular things called ’cherubim’ and a particular relation called ‘typosynthesis’, but rather as simply giving us (or characterizing) a type of structure.
Structures
What is a structure? Let’s start with what is more or less the standard explanation of the concept of structure by saying that a structure is a domain of objects, with one or more relations o n that domain. What needs to be emphasized from the very the beginning is that mathematicians and logicians frequently give specifications of structures that are abstract and general, leaving out of their specifications any mention of the particular objects in the domain of the structure. Thus, the practice is such that:
24
(1) One does not have to say what the things are that are in the domain of the structure. They can be regarded as simply “points” or “p1aces”in the structure, to be filled or replaced by actual or possible things to yield an “instance” of the structure. (2) One does not have to say how the things in the relations are related to one another, more specifically, one does not have to explain what properties things must have or what the things must do, or what spatial or temporal relationships the things must bear to one another, to be in these relations. Thus, from the perspective of set theory, one can specify the (binary) relations of a structure by simply giving the relations as sets of ordered pairs.
I shall understand a mathematician following such a practice t o be specifying a type of structure that specific structures may exemplify or instantiate. Let us reconsider the typosynthesis theory, but this time regard it as simply characterizing a type of structure. In other words, think of the typosynthesis theory in the way Hilbert regarded his geometry. The type of structure in question will consists of all the living humans (henceforth, labeled ‘ W )plus ten other objects, which we can label ‘C’(for cherubim). Then there is a relation R that is such that: (a) For every x that is an H , there is both a y that is a C such that xRy and also a z that is a C such that -xRz. (b) For every x and y, if xRy,then x is an H and y is a C. Now if one thinks of the axioms of the typosynthesis theory as just describing a type of structure, then one will feel no need to answer such questions as “Is the relation R an internal or external relation?” Not being able t o answer such questions does not indicate a lack of understanding of the theory, since the letter “R”is not supposed to stand for a specific relation. Consider now the case of set theory. The above considerations are upsetting t o Lewis and van Inwagen because, as they understand set theory, it is a theory about real objects in the way biology is a theory about real things. They have a view of set theory that is similar t o the view put forward by Godel in his paper (Godel, 1964). This is the famous paper in which Godel responded to the claim that, were the Continuum Hypothesis proved t o be independent of the standard axioms of set theory, the question of the truth or falsity of the hypothesis would lose its meaning, just as the question of the truth or falsity of the Fifth Postulate of Euclidean geometry lost its meaning with the discovery of its independence from the other pos-
25
tulates of Euclidean Geometry.” Godel objected that such an independence result in set theory would render the question of the truth or falsity of the Continuum Hypothesis meaningless, only if set theory were regarded as a hypothetico-deductive system in which the meanings of the primitives of set theory are left undetermined ((Godel, 1964), p. 271). But, for Godel, set theory is not that sort of system. Godel makes it clear that he regards the objects of set theory as things that “exist independently of our constructions”; that we have “an intuition of them individually” (an intuition that is something like a “perception”of individual sets); and that the general mathematical concepts we employ in set theory are “sufficiently clear for us to be able to recognize their soundness and the truth of the axioms c0ncerning”these objects ((Godel, 1964), p. 262). Thus, he declares that “the set-theoretical concepts and theorems describe some well-determined reality, in which Cantor’s conjecture [the continuum hypothesis] must either be true or false”((Gode1, 1964), p. 263-4). For Lewis and van Inwagen, then, the membership relation is like the relation of being taller than: whether someone is taller than someone else is a determinate question of fact, to be determined by going out into the world and measuring. So for these philosophers, the questions as to whether the membership relation is intrinsic or extrinsic, internal or external, are genuine questions about a relation that obtains among real objects. Similarly, the question as to whether the membership relation obtains because of the properties of the things that are in the relation or because of what these things may have done or because of the spatial or temporal relationships that these things have to one another is a genuine questions about a relation that obtains among real objects: for Lewis, these questions must have answers, even if we can’t answer them. The fact that we can’t answer them strongly suggests, to Lewis and van Inwagen, that our grasp of this relation is fundamentally flawed. But think of set theory in the way we regarded the typosynthesis theory: think of the axioms of set theory as merely specifications or characterizations of a type of structure. According to this conception, there will be some domain of objects to be called ‘sets’, and a relation among these objects to be called ’the membership relationship’. The axioms tell us what “Cf. “Probably we shall have in the future essentially different intuitive notions of sets just as we have different notions of space, and will base our discussions of sets on axioms which correspond to the kind of sets we wish to study . . . everything in the recent work on foundations of set theory points toward the situation which I just described”((Mostowski, 1967), p. 94).
26
(so-called) ‘sets’ there must be in the domain and also how the things in the domain must be related by this relation in order that we have a structure of the type in question. Return t o the criteria of understanding I described earlier. You will recall that, in our universities, to determine if someone understands set theory, we see if t h e person can explain t h e principal concepts of t h e theory (by giving t h e relevant definitions, applying these definitions within t h e theory, and explaining their implications for t h e theory), knows t h e axioms or fundamen-
tal assumptions of the area, and also can cite, prove, explain and apply (in a variety of set theoretical contexts) t h e principal theorems of t h e area (both basic and advanced).
These criteria are sensible and appropriate, if we regard set theory in the way I have been suggesting. These criteria test the candidates understanding of the type of structure being characterized and his/her grasp of the detailed information that has been developed about this type of structure. What we do not demand of the candidate is a grasp of the intrinsic properties of sets or an ability to classify the membership relation as internal, external or extrinsic. Thus, if we regard set theory in the way being suggested above, there is no reason why we should expect anyone with the sort of structural understanding of set theory tested by the above criteria to be able to answer the sort of philosophical questions that van Inwagen and Lewis pose. This result, it seems to me, gives us good reason for thinking that the above structural way of regarding set theory is both fruitful and fundamentally sound. Consider now the so-called problem of multiple reductions of number theory that Paul Benacerraf made so famous. For those readers who are unfamiliar with his paper (Benacerraf, 1965), I will give the briefest of sketches of the basic points he makes that are relevant here. Benacerraf describes two imaginary sons of famous mathematicians: one a son of Zermelo and the other a son of von Neumann. The former takes the natural numbers to be the Zermelo finite ordinals; the latter takes the natural numbers to be OCf. Ian Mueller’s comments in (Mueller, 1981), p. 10: “One specifies the structure under consideration by specifying the conditions which it fulfills, i.e., by giving the axioms which determine it. In some cases, the axioms are the only characterization of the structure. For example, in algebra, a group is defined to be any system of objects satisfying certain axioms. In other cases the specification of axioms is an attempt to characterize precisely a roughly grasped structure. Peano’s axiomatization of arithmetic and Hilbert’s Grundlagen are examples.”
27
the von Neumann finite ordinals. Each has been taught by his own illustrious father t o believe that, unlike the vast majority of human beings, he truly knows what the natural numbers are. Benacerraf describes them as arguing over who is right. Of course, neither can give any decisive reason for picking his father’s favored sets as the natural numbers. The moral that Benacerraf wants us to draw from this story is that there is no correct answer to the question. Both of the children’s sets will do equally well as the natural numbers, but so also will countless other omega-sequences of sets. Now from the point of view being set forth here, that is to be expected. If we regard the Peano axioms of number theory as merely specifying a type of structure, then it is by no means surprising that we can model this kind of structure in countless ways in set theory. Similar considerations can lead to the view that the standard axioms of the real number system specify a type of structure. And again, we get multiple reductions of the real numbers to sets of various sorts. For example, the real numbers can be taken to be Dedekind cuts of rational numbers or they can be taken to be Cauchy sequences of rationals. In this case, as in the case of the natural number system, we have a type of structure being modeled in different ways in set theory. No surprise. Thus far, I have been claiming that there is a way of understanding the axioms of mathematical theories, such as set theory, according to which the axioms specify or characterize a type of structure. One attraction of this way of understanding mathematics is that it provides us with a nice way of dissolving and explaining the van Inwagen puzzle. I should add, however, that I have not been claiming that the assertions of the theorythe theorems themselves-make assertions about this type of structure. Nor have I been offering any sort of analysis or translation of the theorems of any mathematical theory. If the point I am making here is not entirely clear to you at this point, do not worry, since I shall be amplifying the point shortly.
The fourth puzzle Let us now consider the fourth puzzle, from the perspective of the structural view of mathematics sketched above. We know that there are many mathematicians who do not believe in the existence of mathematical objects and yet continue t o do fruitful work in mathematics. Is there something about the nature of mathematics itself that does not block or even discourage
28
such ontologically skeptical mathematicians from doing fruitful work in the field? Well what does a mathematician working in, say, set theory do? Typically, the mathematician proves theorems of set theory. So we should ponder the sort of information a theorem of set theory provides to someone who regards the axioms of set theory as specifying conditions that any structure of the type being studied must satisfy. If we regard the axioms of set theory as specifying what (so-called) ‘sets’ there must be in the domain and also how the things in the domain must be related to each other in order that there be a structure of the type in question, then a proof of a theorem of set theory ensures us that any structure satisfying these conditions must also satisfy the conditions given by the theorem. To put it another way, a proof of a theorem provides us with further information about what must hold in any structure satisfying the axioms, and it does so regardless of what literal meaning the theorem may have. A similar conclusion can be arrived at by considering the situation from the perspective of contemporary logical theory. Thus, suppose, for specificity, that the set theory in question is the familiar Zermelo-Fraenkel set theory - a theory formulated in the language of first-order logic. Then any structure that satisfies the axioms of the set theory must also satisfy any sentence of the theory that is derived from the axioms and hence must satisfy any theorem of the set theory. As Ian Mueller puts it, The rules of logic permit the derivation from an axiom system of exactly those assertions that are true under all the interpretations under which the axioms are true. In other words, logical derivations simply bring to light features of the structure characterized by the axioms. ((Mueller, 1981), p. 10).
As I pointed out earlier, I am not here providing an analysis of the literal meaning of any sentence of set theory. Godel believed that our set-theoretical theorems literally describe some well-determined reality, in which Cantor’s conjecture [the continuum hypothesis] must either be true or false. I certainly do not wish to contest Godel’s understanding of what any sentence of set theory literally asserts. Nothing in my explanations of these puzzles requires that I provide a linguistic analysis of the sentences of set theory or of any other mathematical theory. Consider, what a proof of a theorem establishes in the case of Peano arithmetic. There are philosophers of mathematics (e.g. Mark Steiner) who believe that statements of arithmetic are statements about specific abstract objects. As in the case of set theory, I shall not contradict such philosophers
29
or attempt any type of linguistic analysis of number theoretic statements. + 2 = 4’ means, when it is asserted by a mathematician, whether she is a Mathematical Platonist, a formalist, a logicist, or an Intuitionist. I certainly do not claim to know what a school 2 = 4’. Still, there is a way of understanding child may mean by ‘2 the primitive symbols of the theory relative to any structure that satisfies the axioms: the quantifiers can be taken to have as domain the totality of objects in the structure, the individual constants can be taken to refer to specific objects in the structure, and the operation symbols can be taken to refer to the operations specified by the axioms. In this way, we can regard a proof of a theorem in this system as ensuring us that any structure satisfying the axioms of Peano arithmetic must also satisfy the conditions given by the theorem. Thus, the proof of the Commutative Law of Addition ensures us that any domain of objects that are so related as to satisfy the axioms of Peano arithmetic must also be such that, when the primitive symbols are understood in the way described above, the Commutative Law of Addition will also be satisfied. In this respect, the situation is no different in the case of set theory. So (in summary) how do I explain the fact that, in mathematics, there are so many mathematicians who continue to work in set theory, proving important theorems and making important break-throughs, even though they do not believe that any sets exist; whereas, in the empirical sciences, when a scientist does not believe in the existence of the theoretical entities postulated in a particular scientific theory, then typically this scientist will not believe in the theory, work with the theory, perform experiments based upon the theory, and develop explanations in terms of the theory? We can now see an important difference in the kind of theories that are being developed in the two contrasting sciences. In the case of the empirical sciences, the sentences expressing the laws, principles, and known facts of the science are typically taken to be literally true, approximately true, or at least true as some sort of idealization. Furthermore, these laws, principles, and known facts are what get confirmed, applied, and developed. In the mathematical case, the truth of the sentence proved is not what is crucial, because, independently of the literal meaning of the sentence proved, what is established by a proof of a theorem is a truth of the form: any structure satisfying the axioms, or that is of the sort being studied, must satisfy the conditions given by the theorem. From my perspective, it is this information that is most important. Indeed, I would argue that this structural information is all that is required to apply mathematics is the
I do not claim to know what ‘2
+
30
usual ways. Since proofs in mathematics will generate the sort of knowledge that is of genuine interest and value to mathematicians and empirical scientists, regardless of whether such postulated entities as sets, numbers, functions, and spaces actually exist, the value of a mathematician’s work is not dependent upon the truth of the Platonic view. Thus, it is reasonable for a mathematician who is skeptical of the actual existence of, say, sets to work in set theory and to prove theorems in that area.
T h e third puzzle The third of the puzzles to be explained in this work looks back at the history of mathematics and finds a string of brilliant mathematicians making roughly the same basic point, viz. that mathematical existence is a matter of consistency. Consider again the quotation from Hilbert. What the great mathematician had in mind will be clearer, I believe, if we replace the word ‘concept’ that appears in the quotation with the phrase ‘type of object’. Then we have him writing: If contradictory attributes are assigned to a type of object, I say, that mathematically the type of object does not exist. So, for example, a real number whose square is -1 does not exist mathematically. But if it can be proved that the attributes assigned to the type of object can never lead to a contradiction by the application of a finite number of logical inferences, I say that the mathematical existence of the type of object (for example, of a number or a function which satisfies certain conditions) is thereby proved. In the case before us, where we are concerned with the axioms of real numbers in arithmetic, the proof of the consistency of the axioms is a t the same time the proof of the mathematical existence of the complete system of real numbers or of the continuum.
Now why should a proof of the consistency of the axioms of the real number system be tantamount to a proof of the mathematical existence of the system of real numbers? Let us investigate this question within the framework of first-order logic. Let us suppose, in other words, that the axioms Hilbert is writing about were formulated in first-order logic. We know, from a theorem of first-order logic, that a proof of the syntactic consistency of a set of axioms (which is the kind Hilbert seems to have in mind in the quotep) implies that the set of axioms is semantically consistent and hence PNote, especially, Hilbert’s words: ‘ I . . . if it can be proved that the attributes assigned to the type of object can never lead to a contradiction by the application of a finite
31
that there could be a structure of the kind specified by the axioms. In other words, the consistency proof establishes that the mathematical theory of real numbers concerns a genuinely possible kind of structure. This shows us that the theorems of such a system will result in the kind of truths that are of interest to mathematicians, viz. truths that tell us what would have to be the case in structures of the type in question. Thus, the consistency proof would give us a sort of guarantee of the real number system’s mathematical legitimacy, and thus allow us to take the existential statements about the real numbers to be mathematically significant. In this way, we can make good sense of the quotes from Hi1bert.q Similarly, one can understand PoincarB’s views on the same basic model, especially, taking into account his dispute with Bertrand Russell on the nature of geometric axioms.‘ And it is easy to see how thoughts roughly along these lines could have underlain Cantor’s attitudes towards the consistency and validity of his theory of transfinite numbers. A question may arise in the mind of some readers as to how one can reconcile the structural attitude towards mathematics exhibited by Hilbert with the conviction of some mathematicians (such as Kurt Godel) that the assertions of such mathematical theories as set theory concern not structures but genuine mathematical entities (that exist in the real world). To clarify the situation with which we are concerned, I shall make use of a kind of “interpretation”of logical languages that philosophically trained logicians are apt to consider when “translations” of the logical language into some natural language are seriously contemplated. These “interpretations” not only assign the relevant sort of sets and objects to the parameters of the logical language in question, they also supply meanings or senses to the parameters. For example, in (Mates, 1972), “interpretations” of this sort assign to each individual constants the sense of some English name or definite description; and they provide each predicate of the language with the sense or meaning of an English predicate, where English predicates are obtained from English declarative sentences by replacing occurrences of
number of logical inferences . . . ” q 0 f course, such a result linking syntactic consistency with model theoretic consistency cannot be obtained in second-order logic, but we cannot expect Hilbert to have known this, given that the very distinction between first and second order logic had not been made when Hilbert was writing on these matters. ‘See Shapiro’s detailed discussion of the Poincar6-Russell dispute in his (Shapiro, 1997), pp. 153-7.
32
names or definite descriptions with occurrences of circled numeral^.^ I call interpretations of this sort ‘natural language interpretations’ or ‘NL interpretations’ for short.t What is useful, for my purposes, about these NL interpreted languages is that the sentences of such a language can be regarded as expressing statements that are true or false (and not merely true or false in a structure). An example
If t h e interpretation I assigns t o t h e predicate ‘R2’ t h e sense of the English predicate ‘Q is taller than Q’, and t o the individual constants ‘a’ and ‘b’ t h e senses of ‘The vice president of the United States’and ‘The governor of Texas’ respectively, then the sentence ‘R2ba’ expresses the statement ‘The vice president of t h e United States is taller t h a n the governor of Texas’. And that statement is simply true-true of t h e real world.
Now suppose that we have a mathematical theory formalized in the firstorder predicate calculus that is given such an N L interpretation. Hilbert can be regarded as taking the sentences of the theory as formal (uninterpreted) sentences that are true or false only in a structure. In other words, Hilbert can be seen to be interested in the models of the theory and not concerned with the literal meaning of the assertions of the theory. Godel, on the other hand, can be regarded as taking the sentences of the theory as N L interpreted sentences that are just true or false. Thus, from Godel’s perspective, the mathematical theory is like a scientific theory, to be treated as expressing proposition about reality: what is important to Godel is the truth or falsity of the sentences and hence the objective reality to which the sentences are answerable. From Hilbert’s perspective, the axioms of the theory determine a class of structures-the models of the theory-and this is so, independently of the meaning or truth of the sentences. SActually, Mates requires that the occurrences be “direct occurrences”, in order to avoid having the circled numerals occur within the scope of terms of psychological attitude or modal operators, but it is not necessary, for our purposes, to go into these complications. See p. 77 for Mates’ introduction to English predicates, which he tells us he borrowed from Quine. See for example (Quine, 1959), pp. 131 - 134. However, it should be noted that Quine used the expression ’predicate’ to refer to what I am here calling an “English predicate”. I follow Mates both in calling these linguistic items ”English predicates” in order to distinguish them from the predicates of the formal language. I also follow Mates in treating English predicates as devices used in giving a kind of natural language interpretation to formal languages. See (Mates, 1972), pp. 77 - 86, for a discussion of English predicates that more accurately gives the point of view of this work. tQuine discusses such NL interpretations in (Quine, 1959), section 18.
33
T h e second puzzle The inertness feature attributed to mathematical objects gives rise to the questions: (1) How are we able to refer to these inert mathematical objects? (2) How are we able to gain knowledge of these inert objects? These are truly puzzling questions if one assumes, as most contemporary philosophers of mathematics do, that (a) our mathematical theorems express true propositions; and (b) our mathematical theorems are about (make reference to) these inert objects from which we are forever utterly cut off (causally). The structural view put forward in this work provides us with a nice way out of this conundrum, since from the perspective of this view, we do not need to maintain that most mathematical theorems are true or that mathematical theorems refer to causally inert objects. My view is that one of the principal kinds of knowledge mathematics delivers is the knowledge that any structure that satisfies the characterization given by the axioms or that is of the sort being researched must satisfy the conditions given by a theorem when it is interpreted in the structural way discussed earlier . . . . And there is no special problem of how we gain such knowledge: this is just knowledge obtained by means of mathematical proofs. Furthermore, we obtain an explanation of “the primacy of proof as the avenue to mathematical knowledge par excellence”-which Geoffrey Hellman feels “must receive a natural explanation, even if other avenues are left open”((Hellman, 1989), p. 4).
T h e f i r s t puzzle
To understand what is going on in this puzzle, we need t o take account of the enormous gulf that separates Euclid’s conception of geometry from Hilbert’s. Euclid regarded geometry as a theory of physical space: its postulates were supposed to be self-evident truths about space. Hilbert, on the other hand, regarded geometry as a branch of pure mathematics, to be developed from axioms that would characterize a kind of structure. The theorems of Hilbert’s geometry were not only not true of physical space, they were not true at all-at least in the usual sense of that term. This structural conception of geometry did not arise in Hilbert’s mind in complete isolation from the outside influences of the ideas and pronouncements of other mathematicians. Hans F’reudenthal has described the hotbed of ideas in Nineteenth Century mathematics, from which Hilbert’s conception of a new foundations for geometry arose (Freudenthal, 1962). However, the seeds of this structural view of geometry were sewn hundreds of years
34
earlier. At some period, and I won’t speculate exactly when, but surely by the 17th Century birth and development of analytic geometry, especially starting with the work of Fermat, it became commonplace to regard the lines and curves discussed in geometry to be existing in space, independently of our constructions. Such a view of geometry is, of course, quite a departure from the Euclidean conception. Giorgio Israel has recently amplified this point in the following way: Fermat p u t forward, explicitly, t h e principle of t h e one-one correspondence between algebra and geometry, in allowing that, by means of an algebraic equation, i t is possible t o give a geometric locus. The centrality of geometric constructions is eliminated in a single shot. To be admissible, i t is no longer necessary that t h e curve be constructible. T h e curve exists uniquely because its equation is given. It is defined not by means of a construction, but as the locus of points that satisfy the equation. ((Israel, 1998), p. 202).
Thus, mathematicians began to represent lines and curves in space by means of algebraic equations, representing physical space itself as having a mathematical structure which we would now describe as a structure isomorphic to the set of all ordered triples of real numbers, ordered in the familiar way. Under this representation, a position in this space corresponds to an ordered triple of real numbers, a line in space would be a set of such ordered triples that satisfies certain equations, and the space itself would correspond to the totality of all such ordered triples of real numbers. Thus, physical space-the space about which geometers had been theorizing-was given a mathematically definite structural characterization that could eventually lead to an axiomatization of such a structure. It should be mentioned that this structural representation of space is linked conceptually to an array of methods of measurement and comparison of length of physical lines, as well as procedures for carrying out geometric construction in physical space. “The above is a loose translation of the following passage: “Fermat avance, de rnaniere plut6t explicite, le principe de la correspondance biunivoque entre algebra et gBombtrie, en admettant que c’est ii partir d’une equation alg6brique qu’il est possible de donner un lieu g6orn6trique. La centralit6 des constructions g6ornktriques est Blirninke d’un seul coup. Pour Btre admissible, il n’est plus necessaire que la courbe soit constructible, . . . La courbe existe uniquement parce que son Bquation est donnBe. Elle est d6finie non pas au moyen d’une construction, mais cornme le lieu des points qui satisfont 6 I’e‘quation.” Thus, although RenB Descartes is widely thought to be the originator of analytic geometry, there are good grounds for crediting Ferrnat with making the decisive move that gave rise to this new branch of mathematics. For more on this aspect of Ferrnat’s contributions to mathematics, see (Boyer, 1985), pp. 380-2.
35
To the mathematician, what was mathematically significant about physical space were the structural properties of space. So it was natural that, over the years, what became the subject of geometry was the kind of structure that was being attributed to physical space. From this point, it is just a short step to Hilbert’s structural view of geometry. Let us now refocus on the first puzzle. In particular, let us take up the question: How could Hilbert have felt justified in postulating axioms which assert that there exist an infinity of points and lines, when Euclidean plane geometry was developed and applied for a multitude of centuries without any such commitment to an apparent ontology of imperceptible objects?
We can now see that Hilbert’s axioms, when properly understood, do not, in fact, make any such existential assertions. Indeed, we have seen that his geometrical axioms are not assertions at all. They are like the uninterpreted sentences of a first order theory that, in effect, characterize a kind of structure. Hence, in claiming that his axioms express “facts basic to our intuition”, Hilbert was not maintaining that his axioms are assertions about the contents of physical space - that they assert that geometrical points and lines in fact exist. Let me end this paper by distinguishing the view of mathematics I have presented here with a view that has gained much publicity lately in the American literature-mathematical structuralism. This is a view that has been attributed to Paul Benacerraf and Geoffrey Hellman, and is explicitly championed by Michael Resnik and Stewart Shapiro. My view differs from structuralism in several respects, but I do not have the time to go into details. Here is one respect in which my account differs from that of the structuralist. My own view of mathematics does not attempt to provide an account of the content of mathematical assertions in the way that structuralist accounts do: Thus, I do not attempt to describe what “reference to mathematical objects”in typical mathematical theories, such as number theory or set theory, consists in. My position is that, regardless of what may be actually asserted by mathematical theories, there is a way of understanding (or interpreting) the assertions of these theories according to which these assertions are structural in content (every assertions tells us what would have to be the case in structures of a certain type). My structural account of how the assertions of mathematical theories can be interpreted should in no way be taken to be an account of what mathematical theories in fact assert. The claim that mathematical theorems can be
36
interpreted t o be asserting such and such is, of course, significantly weaker than the claim that mathematical theorems in fact assert such and such. This can be seen in the fact that my claims do not require an appeal to detailed empirical studies of the linguistic practices of mathematicians in order to be confirmed or justified, as do the structuralist’s claims. Thus, my position is compatible with an analysis of the content of mathematical sentences that is platonic, nominalistic, or even structuralist. For this reason, I do not classify my view of mathematics as a form of “structuralism”. I prefer, instead, to refer to my view as “a structural account of mathematics”, thereby distinguishing my views from those of such structuralists as Shapiro and Resnik.’
Bibliography Azzouni, J. (1994). Metaphysical Myths, Mathematical Practice. (Cambridge: Cambridge University Press). Benacerraf, P. (1965). What Numbers Could Not Be. The Philosophical Review, 74, 47-73. Bernays, P. (1950). Mathematische Existenz und Widerspruchsfreiheit, ktudes de Philosophie des sciences en hommage d Ferdinand Gonseth, (pp. 11-25). (Neuchatel: Editions du Griffon). Boyer, C. B. (1985). A History of Mathematics. (Princeton: Princeton University Press). Calinger, R. (Ed.). (1982). Classics of Mathematics. (Oak Park, Illinois: Moore Publishing Company). Chihara, C. S. (1990). Constructibility and Mathematical Existence. (Oxford: Oxford University Press). Chihara, C. S. (1998). The Worlds of Possibility: Modal Realism and the Semantics of Modal Logic. (Oxford: Oxford University Press). Cohen, P. (1971). Comments on the Foundations of Set Theory. In Dana Scott (Ed.), Axiomatic Set Theory, (pp. 9-15). (Providence, Rhode Island: American Mathematical Society). Dauben, J. (1990). George Cantor: His Mathematics and Philosophy of the Infinite. (First Paperback printing ed.). (Princeton: Princeton University Press). ”I am preparing a much longer work in which this structural account of mathematics will be presented in great detail.
37
F’reudenthal, H. (1962). The Main Trends in the Foundations of Geometry in the 19th Century. In Ernest Nagel, Patrick Suppes, & Alfred Tarski (Eds.), Logic, Methodology and Philosophy of Science: Proceedings of the 1960 International Congress, (pp. 613-621). (Stanford: Stanford University Press). Godel, K. (1964). What is Cantor’s Continuum Problem? In P. Benacerraf & H. Putnam (Eds.), Philosophy of Mathematics: Selected Readings, (pp. 258-273). (Englewood Cliffs, NJ: Prentice-Hall). Heath, T. (1956). The Thirteen Books of Euclid’s Elements. (Second ed.). (Vol. 1). (New York: Dover). Hellman, G. (1989). Mathematics Without Numbers. (Oxford: Oxford University Press). Hilbert, D. (1971). Foundations of Geometry (Unger, Leo, Trans.). (Second English Translation ed.). (La Salle, Illinois: Open Court). Israel, G. (1998). Des Regulaeii la Gkomktrie. Revue d’Histoire des Sciences, 51, 183-236. Kogbetliantz, E. G. (1969). Fundamentals of Mathematics from an Advanced Viewpoint. (Vol. 3). (New York: Gordon and Breach Science Publishers). Lear, J. (1977). Sets and Semantics. The Journal of Philosophy, 74, 86-102. Lewis, D. (1983). New Work for a Theory of Universals. Australasian Journal Of Philosophy, 61, 343 - 377. Lewis, D. (1991). Parts of Classes. (Oxford: Basil Blackwell). Maddy, P. (1980). Perception and Mathematical Intuition. The Philosophical Review, 89, 163-196. Maddy, P. (1990). Realism in Mathematics. (Oxford: Oxford University Press). Mates, B. (1972). Elementary Logic. (2nd ed.). (New York: Oxford University Press). Mostowski, A. (1967). Recent Results in Set Theory. In I. Lakatos (Ed.), Problems in the Philosophy of Mathematics, (pp. 82-96). (Amsterdam: North-Holland) . Mueller, I. (1981). Philosophy of Mathematics and Deductive Structure in Euclid’s Elements. (Cambridge, Massachusetts: M. I. T. Press). Poincark, H. (1952). Science and Hypothesis. (New York: Dover Publications, Inc.).
38
Poincarb, H. (1953). Science and Method (Maitland, Francis, Trans.). (New York: Dover Publications, Inc.). Quine, W. (1959). Methods of Logic. (Revised Edition ed.). (New York: Henry Holt & Company, Inc.). Robinson, A. (1965). Formalism 64. In Yehoshua Bar-Hillel (Ed.), Logic, Methodology and Philosophy of Science, (pp. 228-246). (Amsterdam: North-Holland Publishing Company). Shapiro, S. (1997). Philosophy of Mathematics: Structure and Ontology. (Oxford: Oxford University Press). van Inwagen, P. (1986). Two Concepts of Possible Worlds. Midwest Studies in Philosophy, 11, 185-213.
Wolff, P. (1963). Breakthroughs in Mathematics. (New York: The New American Library).
39
COMPLEXITY CLASSES OVER THE REALS: A LOGICIAN’S VIEWPOINT *
FELIPE CUCKER Department of Mathematics City University of Hong Kong 83 Tat Chee Avenue, Kowloon HONG KONG enlad: macucker@math. cityu. edu. hk
1. Introduction
A major research topic within mathematical logic in the 1930’s was the quest for a formal notion of computability. Somehow surprisingly, the many answers provided by Church, Godel, Kleene, Markov, Post, and Turing, among others, turned out to be equivalent. Thus Church postulated, in what is known as Church’s Thesis, the equivalence of any other reasonable notion of “computable” with the notions in the mentioned answers. One such notion, the Turing Machine, was going to play a central role in the development of theoretical computer science. Due to its semantic nature -the Turing machine is a theoretical devise upon which a notion of computation is naturally defined- the concepts of running time and workspaceare apparent and, therefore, the grounds of a theory of complexity are easily laid out using this model. This was actually done during the 1960’s and 1970’s and we have today a well-developed theory which classifies thousands of computational problems in a number of complexity classes and draws a web of relationships between the latter. Classes like P, NP, and PSPACE have an outstanding position in this landscape (of which the books [l,2,181 present good pictures). The question of whether P = NP is considered one of the most important open questions in mathematics [19]. The same question for other major complexity classes poses a variety of basic open problems in theoretical computer science. ‘This work has been partially supported by City University grant 7001290
40
An approach which may shed some light on these questions consists on characterizing complexity classes in terms different to their Turing machine based definitions. Mathematical logic has been at the center of this approach providing at least two more ways to look at these classes. Firstly, descriptive complexity. Here, decision problems are seen as sets of finite models satisfying a finite number of sentences. It is then observed that the larger the complexity class, the bigger the expressive power needed to write down the sentences above. For instance, problems in P are those described by sentences in fixed-point first-order logic [15,20] and problems in NP are those described by sentences in existencial second-order logic [12]. Secondly, implicit complexity. Here the inspiration comes from recursion theory, more precisely, from the Godel and Kleene algebraic characterization of recursive functions. This characterization describes the set of recursive functions as the smallest set containing some basic functions and closed under a few operations (composition, recursion, and minimization). In [3], Bellantoni and Cook extended this idea to characterize the class of functions computable in polynomial time. The preceding lines roughly describe a theory (or theories) which apply mainly to discrete structures as they appear in the design and analysis of algorithms in computer science. With the dawn of the computers and apart from the theory above, a different tradition arose around the subject of numerical computations as they are performed in numerical analysis. Here the problems are of an algebraic and analytic nature rather than combinatorial and they consider as inputs finite vectors over a field. A special emphasis is made on the field of the real numbers since this is the case numerical analysis deals with. A cornerstone in this tradition is a paper by L. Blum, M. Shub and S. Smale [5] which introduced a machine model allowing the development of a complexity theory over the reals similar to the one built around the Turing machine, which we already called classical. Real versions of P and N P were defined in [5] and the existence of NPR-complete problems was proved. In the last few years several results were proved for this machine model by a variety of authors. For an overview of this see [4] as well as the survey paper [16]. The machine model introduced in [5] can be seen as a generalization of the Turing machine in which computations can be performed over an arbitrary base ring R. In case R = Q , the field of two elements, we obtain a model which is equivalent to the Turing machine. In this paper we review some of the extensions of descriptive and implicit
41
complexity t o the Complexity theory over the reals introduced by Blum, Shub and Smale. For descriptive complexity, the first such generalization was done by Gradel and Gurevich in [13]. They introduced the notion of metafinite structure in which a finite structure is endowed with a set of functions into another structure, possibly infinite. Our case of interest, when this second structure is R,was studied by Gradel and Meer in [14]. Here they characterize some complexity classes over the reals, such as PIR or NPR, in terms of logics for metafinite structures over IR, called IRstructures. Subsequent work [lo] captured several other complexity classes. For implicit complexity, this generalization is newer and it can be found in [6]. We review descriptive complexity over IR in Section 3 and implicit complexity over IR in Section 4. The next section is intended to recall the reader the basic objects of the theory initiated by Blum, Shub and Smale. 2. Machines and complexity classes over the reals
We denote by R" the disjoint union
R"
= Un>lRn, -
where for n > 0, Rn is the standard n-dimensional space over R. The space IR"O is a natural one to represent problem instances of arbitrarily high dimension. For x E Rnc IR", we call n the size of x and we denote it by size(x). In this paper we will consider BSS-machines over R as they are defined in [5,4]. Roughly speaking, such a machine takes an input from R", performs a number of arithmetic operations and comparisons following a finite list of instructions, and halts returning an element in IR" (or loops forever). For a given machine M , the function ( P M associating its output t o a given input x E IR" is called the input-output function. We shall say that a function f : Roo+ Rk, k 5 cm,is computable when there is a machine M such that f = ( P M . Also, a set A C Roois decided by a machine M if its characteristic function X A : RW+ ( 0 , l ) coincides with ( P M . So, for decision problems we consider machines whose output space is ( 0 , l ) c IR. We next introduce some central complexity classes.
Definition 2.1. A machine M over R is said to work in polynomial time when there are constants c,q E W such that for every input x E IR", M
42
reaches its output node after at most csize(z)q steps. The class PR is then defined as the set of all subsets of IR," that can be accepted by a machine working in polynomial time and the class FPR as the set of functions which can be computed in polynomial time.
Definition 2.2. A set A belongs to NPR if there is a machine M satisfying the following conditioil: for all x, x E A iff there exists y E lR" such that M accepts the input (x,y) within time polynomial in size($). In this case, the element y is called a witness for z. If we require the witness y to belong to (0, 1)" we say that A E DNPR (the D standing for digital). We will sometimes abuse of language and call the machine M above an NPR-machine (resp. a DNPR-machine). Remark 2.1. (i) In this model the element y can be seen as the sequence of guesses used in the "tiring machine model. However, we note that in this definition no nondeterministic machine is introduced as a computational model, and nondeterminism appears here as a new acceptance definition for the deterministic machine. Also, we note that the length of y can be easily bounded by the time bound p(size(z)). (ii) The class NPR can be used as the building block of a hierarchy = PR and, for k 2 0, Ck+' = of complexity classes. Define
ck
NPRR. Also, let hierarchy
llg = {R" \ S I S PHR =
E E g } and the poZynomiaZ
U Cg. k>O
In a similar way one defines the digital arithmetical hierarchy DPHR. An example of a set in NPR is 4-FEAS, the set of polynomials of degree four which have a real root. A polynomial f is considered as an element in IR" by coding it by the sequence of its coefficients. Note that if f has n variables, then it has C?(n4)coefficients. Given such an f and a guess z (which we suppose of size n) a machine deciding 4-FEAS just computes f(z)and accepts if this is zero rejecting otherwise. The set 4-FEAS is also an example of a "difficult" NPR problem in a precise sense. The following is proven in [5] where the definition of NPRcompleteness can also be found.
43
Theorem 2.1. (151) T h e set 4-FEAS is NPR-complete for reductions in P R. 0 Parallelism can also be considered for computations over the reals. We shall now briefly recall a parallel computational model. Let the sign function sign : R
+ {0,1}
be defined by sign (2)= 1 if z 2 0 and 0 otherwise.
Definition 2.3. An algebraic circuit C over R is an acyclic directed graph where each node has indegree 0, 1 or 2. Nodes with indegree 0 are either labeled as input nodes or with elements of R (we shall call them constant nodes). Nodes with indegree 2 are labeled with the binary operators of R, i.e. one of {+, X , -, /}. They are called arithmetic nodes. Nodes with indegree 1 are either sign nodes or output nodes. All the output nodes have outdegree 0. Otherwise, there is no upper bound for the outdegree of the other kind of nodes. Occasionally, the nodes of an algebraic circuit will be called gates. For an algebraic circuit C, the size of C, is the number of gates in C. The depth of C, is the length of the longest path from some input gate to some output gate. Let C be an algebraic circuit with n input gates and m output gates. Then, to each gate g we inductively associate a function fg : R" + IR. We shall refer to the function ipc : Rn -+ R" associated to the output gates as the function computed by the circuit.
Definition 2.4. Let f : IR" + IR". We shall say that the family of algebraic circuits {C}>,, - computes f , when for all n 2 1 the function computed by C, is the restriction of f to Rn C R". We now require a condition on the whole family {Cn},>l in order to ensure that its elements are not too unrelated as well as to ensure a finite description of the machine model. Gates of algebraic circuits can be described with five real numbers in a trivial way (see, e.g. [4]). Therefore, a circuit of size li can then be described by a point in IR5k.
Definition 2.5. A family of circuits {C,},,, is said to be u n i f o r m if there exists a machine M that returns the description of the ith gate of C, with input (n,i). In case that i > Ic, the number of gates of C,, M returns
44
(i,0, 0, 0,O). If M works in time bounded by O(log n) we shall say that the family is L-uniform, if M works in time O ( n k )for some positive integer k we shall say that the family is P-uniform. We now define some parallel complexity classes by bounding the depth and size of uniform families of circuits. For further details about these classes see [4,7,8]. Definition 2.6. Define NCL for Ic 2 1to be the class of sets S C IR" such that there is a L-uniform family of algebraic circuits {C,} having size polynomial in n and depth O(logk n ) that computes the characteristic function of S. The union of the N C k is denoted by NCR. We define PARR to be the class of all sets S C R" such that there is a P-uniform family of algebraic circuits {C,} having depth polynomial (and therefore size exponential) in n that computes the characteristic function of S. Also, we define FPARm to be the class of functions f : R" + R" such that 1 f (.)I = ( x ( ~for ( ~all) x E R" and f can be computed by a P-uniform family of algebraic circuits {C,} having depth polynomial and size exponential in n. 3. Descriptive complexity over R 3.1. Logics o n lR-structures In this section we first recall basic notions of R-structures and their logics. A main reference is [14]where these concepts were first introduced. We suppose the reader familiar with the main terminology of logic as well as with the concepts of vocabulary, first-order formula or sentence, interpretation and structure (see for example [ll]). Definition 3.1. Let L,, L f be finite vocabularies where L, may contain relation and function symbols, and L f contains function symbols only. A R-structure of signature u = (L,, L f ) is a pair 3 = (U, F)consisting of (i) a finite structure U of vocabulary L,, called the skeleton of D, whose universe A will also be said to be the universe of D, and (ii) a finite set F of functions X : Ak + R interpreting the function symbols in L f . We shall denote the set of all R-structures of signature u by StructR(a). Definition 3.2. Let 3 be a IR-structure of skeleton 'u. We denote by IAl the cardinality of the universe A of 'u. A R-structure D = (U,F) is ranked if there is a unary function symbol r E L f whose interpretation p
45
in T bijects A with (0, 1,. . . , IAl - 1). The function p is called ranking. A k-ranking on A is a bijection between Ak and {0,1,. . . , IAlk - 1).
3.2. First-order logic Fix a countable set V = {vo,01,. . . } of variables. These variables range only over the skeleton; we do not use element variables taking values in R.
Definition 3.3. The language FOR contains, for each signature (T = (L,, L f ) a set of formulas and terms. Each term t takes, when interpreted in some R-structure, values in either the skeleton, in which case we call it an index term, or in R,in which case we call it a number term. Terms are defined inductively as follows (i) The set of index terms is the closure of the set V of variables under applications of function symbols of L,. (ii) Any real number is a number term. (iii) If h l , . . . , hk are index terms and X is a k-ary function symbol of L f then X ( h 1 , .. . ,hk) is a number term. (iv) If t ,t' are number terms, then so are t t', t - t', t x t', t/t' and sign ( t ).
+
Atomic formulas are equalities hl = h2 of index terms, equalities tl = t z and inequalities tl < tz of number terms, and expressions P ( h 1 , .. . ,hk) where P is a k-ary predicate symbol in L , and h l , . . . ,hk are index terms. The set of formulas of FOR is the smallest set containing all atomic formulas and which is closed under Boolean connectives and quantification (3v)lc, and (Vv)lc,. Note that we do not consider formulas (3x)lc, where x ranges over R.
Remark 3.1. The interpretation of formulas in FOR on a R-structure 9 is clear. The only remark to be done is that, as with circuits, in order to have this interpretation well defined, we understand that z/O = 0 for all x E R. Example 3.1. Let L, be the empty set and L j be { r , X } where both function symbols have arity 1. Then, a simple class of ranked R-structures with signature (L,, L j ) is obtained by letting U be a finite set A , r9 any ranking on A and XD any unary function X D : A + R. Since r9 bijects A with {0,1,.. . ,n - l} where n = / A [ ,this R-structure is a point XD in IR". Conversely, for each point z E Rm there is an R-structure 9 such that x = XD. Thus, this class of structures models R".
46
On the other hand any R-structure 9 = (U,F)can be identified with a vector e ( 9 ) E IR" using a natural encoding. To this aim choose a ranking on A. Without loss of generality the skeleton of D can be assumed to consist of the plain set A only by replacing all functions and relations in L, by their corresponding characteristic functions -the latter being considered as elements of the set F. Now using the ranking each of the functions X in T can be represented by a vector vx E IR" for some appropriate m. The concatenation of all these vx yields the encoding e ( 9 ) E R". Note that the length of e ( 9 ) is polynomially bounded in IAI; moreover for all R-structures D, all rankings E on A and all functions X : Ak -+ R the property that X represents the encoding e ( 9 ) of 9 with respect to E is first-order expressible (see [14]). Example 3.1 allows us to speak about complexity classes among IRstructures. If S is a set of IR-structures closed under isomorphisms, we say that S belongs to a complexity class C over the reals if the set {e(D) I 9 E S } belongs to C.
Example 3.2. If D is a R-structure of signature ( L s ,L f ) and r E L f is a unary function symbol we can express in first-order logic the requirement that r is interpreted as a ranking in 9. This is done by the sentence
r is injective A 30 r ( o ) = 0 A Vu [u# o +-( r ( o )< r ( u )A 3v r ( u ) = r(v)+l)]. Remark 3.2. If p is a ranking on A and IAl = n then, there are elements 0,1 E A such that p(o) = 0 and p(1) = n - 1. Note that these two elements are first-order definable in the sense that they are the only elements in A satisfying
vv (v #
0
* P(0) <
P(V))
and
hJ (v # 1
*
P(V>
< P(1))
respectively. We shall take advantadge of this property by freely using the symbols o and 1 as symbol constants that are to be interpreted as the first and last elements in A with respect to the ranking p. Note that, in particular, this allows us to use the symbol n to denote the cardinality of A since n = p(1) 1.
+
Remark 3.3. Any ranking p induces, for all k 2 1 a Ic-ranking pk on A by lexicographical ordering. Note that pk is definable in the sense that for all
47
(211,
... , ~ k €) A k p k ( v l , .. . , v k ) = p(vl)n"'
+ p(v2)nk-2+ . . . + p(vk).
Again, we will take advantadge of this to freely use the symbol pk to denote the k-ranking induced by p on A. The expressive power of first-order logic is not too big.
Proposition 3.1. Let u be a signature, cp a first-order sentence and S = {D E StructR(u) I D cp}. Then S E NCh. 3.3. Fixed point first-order logic
A first-order number term F(?) with free variables ? = ( t l ,. . . , t r ) is interpreted in a R-structure 9 with universe A as a function F 9 : A' + R. Fixed point first-order logic enhances first-order logic with the introduction of two grammatical rules to build number terms: the maximization rule and the fixed point rule. The first one, allows some form of quantification for describing F m and the second one, the definition of F9 in an inductive way. For simplicity, in the rest of this paper we restrict attention to functional R-structures, i.e. R-structures whose signatures do not contain relation symbols. This represents no loss of expressive power since we can replace any relation P C Ak by its characteristic function x p : Ak + IR. We first define the maximization rule MAXL.
Definition 3.4. Let F(s,?) be a number term with free variables s and ? = ( t l ,... ,tr). Then max F ( s , f ) S
is also a number term with free variables 3. Its interpretation in any Rstructure 9 and for any point u E A' interpreting ? is the maximum of F9 ( a ,u)where a ranges over A.
Example 3.3. If the signature contains a symbol r which is interpreted as a ranking, then we can define the size n of the universe with the number term maxr(s) 1. S
+
Definition 3.5. We denote by FOm+MAXk the logic obtained by adding to FOR the maximization rule.
48
The expressive power gained by allowing the maximization rule lies in the possibility of writing characteristic functions as number terms. If cp(w1,. . . , v,) is a first-order formula we define its characteristic function ~ [ c p ]on a structure 9 by
where a l , . . . ,aT E A, the universe of D.
Proposition 3.2. For every first-order formula number term in FOR MAX:, describing ~ [ c p ] .
+
cp(w1,.
. . , v,) there is a
We now define the fixed point rule.
Definition 3.6. Fix a signature c = ( L , , L f ) , an integer T 2 1, and a pair ( 2 , D ) of function symbols both of arity T and not contained in this signature. Let F ( 2 ,f) and H ( D ,f) be number terms of signature ( L s ,L f U (2,D } ) and free variables t = (tl,. . . , t T ) .Note that 2 can appear several times in F and we do not require that its arguments are t l , . . . ,t,. The only restriction is that the number of free variables in F coincides with the arity of 2. A similar remark holds for H and D. For any R-structure D of signature (T and any interpretation : A' + R of 2 and A : A' + R of D respectively the number terms F ( 2 , f ) and H ( D ,?) define functions
<
F F , HE : A'
+ R.
Let us consider the sequence of pairs {Ai,C}i>o - with Ai : A' + R inductively defined by
A'(%)= 0
for all z E A'
('(x) = 0
for all x E A'
Ai+'(x) =
ci : A' + IR and
H g i ( z ) if Ai(x) = 0
A i ( z ) otherwise Ci+'(.)
=
F;(z) if Ai(x) = 0
c ( x ) otherwise. Since Ai+'(x) only differs from A i ( z )in case the latter is zero, one has that Aj = Aj+' for some j < IAI'. In this case, moreover, we also have that < j = <j+'. We denote these fixed points by 2" and D" and call them the fixed points of F ( 2 , S ) and H(D,T) on D. We say that F D updates C.
49
Note that D plays the role of the characteristic function for the domain of Z and the different A idetermine the successive updatings of this domain. We say that 2" is defined on x if D"(x) # 0. The fixed point rule is now stated in the following way. If F ( Z , t ) and H ( D ,t ) are number terms as above then
fP[Z(f) + F(Z,t),H(D,f)l(q and f P [ m + H ( D ,w4
are number terms of signature ( L s ,L f ) . Their interpretations on a given R-structure 2 are Z"(U) and D" (21) respectively.
A simple example of a function definable with the fixed point rule, which we will use in the next section, is the exponential. Example 3.4. Consider a signature of ranked R-structures and let r be the symbol for the ranking. We will define a function 2T by means of the above fixed point rule such that for all ranked lR-structure 9 the interpretation of 2' is 2p : A -+ (0,.
. . , 2n-1}
x -+ 24") where A is the universe of 9, n = IAl and p is the ranking which interprets r.
To do so, consider the number terms F ( 2 ,x)
+
x[r(z)= 01 + maxx[r(x) = r ( s ) 1]2Z(s) s and H ( D ,x) x[r(x) = 01
+ maxX[r(x) = r ( s ) + 1]D(s). s
One can check that 2" is the function 2f described above. In a similar way, for Ic 2 2 one defines a function 2Tk which is interpreted as 2Pk : A -+ (0,. . . ,2nk-1} x -+ 2 f k W
Definition 3.7. Fixed point logic for R-structures, denoted FPL, is obtained by augmenting first-order logic FOR with the maximization rule and the fixed point rule.
50
Fixed point logic captures polynomial time.
Theorem 3.1. ([14]) Let 0 be a signature and S be a decission problem of ranked R-structures over the sagnature 0. Then the following two statements are equivalent. (2) s E PR (ii) there exists a sentence 9
I= II,}.
II, in FPL such that S = {9 E StructR(0) I 0
Remark 3.4. We will abreviate in the sequel a statement like the one of Theorem 3.1 in the following way. For ranked R-structures, FPL = PR. The classes NC& can be characterized with some extensions of firstorder logic. Since we already know that fixed point first-order logic captures the class PR, we shall look at some logic between FOR and F P L . We shall do so by limiting the number of updatings in the definition of fixed points.
Definition 3.8. Let J : N + N be a function. The formulas of f-bounded fixed point logic FP&[f]are those in FPL. The interpretation of a formula cp in a R-structure !B of size n is as in F P L with only one difference; while defining a function Z as a fixed point, the updating procedure performs at most f (n)steps. If after this number of steps the fixed point has not been reached, the f (n)th update of the function is taken as the interpretation of 2.
If F' is a set of functions f : IN
+ N we denote by FPk[.F] the union
u
FPL[fl.
fG=
For instance FPLIL?(logkn)]=
U FPL[c(logk n ) + 4. c,dEN
Theorem 3.2. For all k 2 1, FPk[B(logk n)] = NC&.
3.4. Complexity classes beyond PR The catalog of complexity classes over the reals is not so vast as in the classical case. Nevertheless, it includes several natural complexity classes. For instance PR and EXPR are the real versions for polynomial and exponential
51
time respectively. Also, NPR is the class of sets decided in nondeterministic polynomial time, and the possibility of alternating a bounded number of existential and universal guesses leads to the polynomial hierarchy PHR over the reals. A difference which stands out is the power of space in the real and classical settings. While in the latter one has that PH PSPACE g EXP, in the former it has been proved in [17] that every decidable set can be decided using a polynomial number of registers. Yet we can consider the class PARR defined in Section 2 as a real analogue of PSPACE. We can summarize this situation with the diagram
NPm . ..
7
L
L
7
PHm
PR C O N P .. ~.
+ PARR 3 EXPR
where an arrow + means inclusion and an arrow 3 means strict inclusion. In this and the next section we will capture the classes in the diagram above with specific logics on R-structures.
Definition 3.9. We say that $ is an existential second-order sentence (of signature u = ( L 8 ,L f ) )if $ = 3Y1 . . . 3Y,+ where q5 is a first-order sentence in FOR of signature ( L s ,L f U . . . ,Y,}). The symbols Y I ,. . . ,Y, will be called function variables. The sentence 1c, is true in a R-structure D of signature u when there exist interpretations of Y1,. . . ,Y, such that 4 holds true on D. The set of existential second-order sentences will be denoted by ISOR. Together with the interpretation above it constitutes existential second-order logic.
{x,
Example 3.5. ([14]) Let us see how to describe 4-FEAS with an existential second-order sentence. Consider the signature (0,{ r , c } ) where the arities of r and c are 1 and 4 respectively, and require that r is interpreted as a ranking. Let D = (U, F)be any R-structure where F consists of interpretations C : A4 -+ R and p : A -+ R of c and r . Let n = [ A ]- 1 so that p bijects A with {0,1,. . . ,n}. Then D defines a homogeneous polynomial $ E IR[Xo,. . . ,Xn] of degree four, namely =
c
(i,j,k,l)EA*
C(i,j,k,l)X,X,X,Xe.
52
We obtain an arbitrary, that is, not necessarily homogeneous, polynomial Xo = 1 in $. We also say that 9 defines g. Notice that for every polynomial g of degree four in n variables there is a R-structure 9 of size n 1 such that 9 defines 9. Denote by 0, 1,is and i the first and last elements of A and A4 with respect to p and p4 respectively. The following sentence quantifies two iand Y : A4 + R functions X : A -R g E R[X1,. . . ,X n ] of degree four by setting
+
111 = (EIX)(EIY)( ~ ( 5=) C ( B ) A y(i)= o A ~ ( 0=)1 A A V U .~. . b ’ ~ 4 [U # 0 * 3 1 . . . 3 ~ 4(p4(u) = p4(w) + 1) A
Y ( u )= Y ( w + ) c(U)X(~i)X(~~)X(ucL3)x(.L1,)1).
Here, if ai = p - l ( i ) for i = 1,... ,n then, ( X ( a l ) , ... ,X(a,)) E R” describes the zero of 9. The latter polynomial is now evaluated monomial by monomial at the point ( X ( a , ) ,. . . ,X(a,)). This is done by letting u cycle through A4 and using Y in order to hold the sum of all monomials of g up to the current u (according to the considered ordering on A4). The update of Y is described by the last row of the formula. The sentence II, describes 4-FEAS in the sense that for any R-structure 9 one has that 9 111 if and only if the polynomial g of degree four defined by B has a real zero.
+
The fact that existential second order logic describes a NPR-complete problem is not fortuitous. Theorem 3.3. 0141) 3
0 =~ NPR.
0
Denote by SOR the logic obtained by allowing quantifiers in Definition 3.9 to be both universal and existential. We then have the following result. Corollary 3.1. SOR = PHm.
0
Nondeterminism for real machines as introduced in [5]appears as a sequence of “guesses” y1, . . . ,ym where the yi are real numbers. It formalizes the idea of finding a witness in a continuous space and problems like 4FEAS capture the difficulty of such a search. However, this kind of search is not the only possibility. In many situations a witness is found in a discrete space and we only need to guess one among a finite set of possibilities. Let us define this more formally (for further details see [9]).
53
Definition 3.10. A set S Roobelongs to DNPm when it can be decided in polynomial time by a nondeterministic machine whose guesses are restricted to be in the set {0,1). We say that the machine above is a digital nondeterministic machine. Natural examples of problems in DNPm exist. The real versions of the Knapsack and the Travelling Salesman problems belong to DNPm. Theorems 3.1 and 3.3 suggest a logic for capturing DNPR. Definition 3.11. Let (Ls, L f ) be a signature. A symbol Y E L j is said to be digital when for every R-structure its interpretation has image included in ( 0 , l ) . We say that qb is an existential digital second-order sentence (of signature (Ls, L f ) if qb = 3Yl . . .3Y,q5 where q5 is a fixed point first-order sentence of signature ( L s ,L f U {Yl,. . . ,Y,}). The symbols Y1,. . . ,Y, here are digital and will be called subset variables. As above, the set of existential second-order formulas will be denoted by 3DSOm. Digital second-order logic DSOR is obtained by allowing the quantifiers above to be universal as well.
A simple variation of Theorem 3.1 yields the following. Theorem 3.4. 3DSOm = DNPm.
0
3.5. A fixed point rule f o r DSOm The last goal of this section is to capture the classes EXPm and PARm. To do so we will enhance DSOR with some fixed point rule to inductively define functions which take as arguments subsets of the universe as well as its elements. As with first-order logic, we also define a maximization rule to be able to write characteristic functions of formulas in DSOR as number terms. In order to apply these functions we need to define the right class of terms to pass them as arguments. Index terms correspond to functions valued in the universe and number terms to real valued functions. None of them represent subsets of the universe. Subset terms as defined below represent subsets of Ak in a natural way. Definition 3.12. Let T ( x 1 , . . . ,x,) be a number term with free element variables $1, . . . ,2,. For any R-structure D consider the digital function
54
A' (1 '
7 * ' *
,
+ {0,1} +
{
l i f P ( ~ 1... , ,v~)#O 0 otherwise
A subset term is either a digital function symbol X or an expression of the form 1[7(z1,.. . , x r ) ] . The interpretation of the subset term [ ~ ( q . .,,G)] . is the function 7above. We also extend the definition of number term to include expressions of the form F ( z 1 , . . . , z,) where F is a subset term of arity T and X I , .. . ,z, are index terms.
Example 3.6. Some simple examples of subset terms are given by the formulas r
A ~ ( z j<) 0
7 0 ( ~ 1 ,... ,GI =
j=1
which defines the empty set, r
A P(q) 20
m(z1, * . . 7 z r ) =
j=1
which defines the total set T = A', r
7r(zi,...
AP(~~)=o
,zr)=
j=1
which defines the set I = ((0,. . . ,o ) } , r
TM(z1,
... ,xi)= A
p(q)=n- 1
j=1 which defines the set M = { (1,.. . ,l)}, and T M + I ( ~ *I .. , , z r ) = ~ ( z l *, .> z r )v ~ 1 ( 2 1 ,... 7 z r )
+
which defines the set M 1 = ((1,.. . ,l), (0,.. . , o ) } . We shall denote by 0, T, I, M and M + l the subset terms defined by them. We now define the maximization and fixed point rules for digital secondorder logic. The definition closely follows the one for first-order logic done in 53.1.
55
Definition 3.13. Let S and T = ( T I , . . ,T,) be subset variables with arities Ic and a l , . . . ,a, respectively. Let also Z = ( t l ,. . . , t,) be element variables and F ( S ,Z, F ) be a number term having S and T as free subset variables and Z as free element variables. Then
max F ( S ,T, T ) S
is also a number term with free variables 3,T. Its interpretation in any IR-structure 9 for any sets Ui E ?(Aai),i = 1,.. . ,s, interpreting T and any elements 211,. . . ,u, in A interpreting Z is the maximum of F 9 (V, V) where V ranges over ? ( A k ) .
c,
Definition 3.14. We denote by DSOR adding to DSOR the maximization rule.
+ MAX& the logic obtained
by
Again, the maximization rule allows us to write characteristic functions of digital second-order formulas as number terms. Proposition 3.3. For all digital second-orderformula cp there is a number 0 term in DSOR MAX& describing ~ [ c p ] .
+
We now define the fixed point rule. In order to do so we need to use a new kind of signature. Before introducing it, notice that digital function , set symbols can be interpreted in a natural way as elements in P ( A T ) the of parts of A' where r is the arity of the symbol. Definition 3.15. A second-order signature is a triple (L,, L f , Lo) where (Ls, L f ) is a signature as defined in Definition 3.1 and Lo is a finite set of function symbols called second-order functionals. Each of them has associated a tuple (r,s,a l , . . . ,a,) of natural numbers. The number r is the element arity, s is the subset arity, and the numbers a l , . . . , a , are the argument arities. A second-order functional takes as argument tuples of the form (q,... , z T , X 1 , . .. , X , ) where zi is an index term for i = 1,... , r and X i is a subset term of arity ai for i = 1 , . . . ,s. Thus a symbol 2 in Lo is interpreted on a R-structure 9 of universe A as a function
29 : A' x P ( A a l )x . . . x ?(Aaa)
+ IR.
Remark 3.5.
i) A digital function symbol, say X , can occur in a formula in two different ways. This is the case for instance in the term
Z ( X )+ X b )
56
in which Z denotes a second-order functional and x an element variable. ii) We won't consider R-structures over second-order signatures in what follows but only as a technical trick to define the secondorder fixed point rule.
Definition 3.16. Let s, a l , . . . ,a, be natural numbers and Xi be subset variables of arity ai for i = 1,.. . ,s. Let also 21,.. . ,x, be r element variables, r E IN. Denote Z = (XI,.. . ,z,) and = ( X I , .. . ,X,). Fix a signature c = ( L s ,L f ) and a pair ( 2 , D )of second-order functionals both of element arity T , subset arity s and argument arities a l , . . . , a,. Thus 5 = ( L , , L f ,( 2 , D ) ) is a second-order signature. Let F ( 2 , Z , y ) and H ( V , Z , Z ) be number terms of signature (L,,Lf,{Z,D}) which have y as free subset variable and Z as free element variables. For any R-structure 9 of signature (T and interpretations
C, A : A'
x "(A"') x
. . . x "(A".) -+ R
x)and H ( D ,?f,x)define
of 2 and 27 respectively the number terms F ( Z , Z , functions
FF, HE : A' x "(A"') x . . . x P(A".) + R. Consider the sequence of pairs { A i ,Ci}i>0 with P ( A a S )-+ R and A i: A' x P(A"1)x . . . ;"(Aas) by
(k,U)E A'
c
: AT x
F(Aal)x
.. . x
+ R inductively defined
x P(A"') x
Ao(E,U)= 0
for all
for all ( E , u ) E A' x "(A"') x
.. . x P ( A a S ) . . . x "(A".)
Fixed points Zm and V m are defined as in the preceding section. Now however, the number of steps in the inductive definition is bounded by 2ne where n = IAl for a suitable l . Example 3.7. Consider a signature of ranked IR-structures and let T be the symbol for the ranking. We will define a function r2 such that for all
57
-
ranked R-structures D the interpretation of r k is
-
pk : ? ( A k ) + ( 0 , . . . ,2nk - 1}
where A is the universe of D, n = IAl and p is the interpretation of the ranking. To do so, consider the number terms F ( 2 ,X )
+
maxX[3a: x = u u { T } I ( Z ( U ) 2'"(:)) U
and H ( D , X )
x[X = 01
+ maxx[3: U
X = U u {a:}]D(U)
-
where Z = (zl,. . . ,x k ) . One can check that Zoi' is the function p k above. Here we are usin% the exponzntial function aTk-described in Example 3.4. Notice that p k ( 0 ) = 0, pk((T) = 2nk - 1, pk(I) = 1, p k ( M ) = 2 n k - 1 , and p'e(M + 1) = 2nk-1 1 for the sets T , I , M and M 1 defined in Example 3.6.
-
+
+
Definition 3.17. Fixed point digital second-order logic for R-structures, denoted FP&, is obtained by augmenting DSOR with the maximization rule and the fixed point rule described above.
Theorem 3.5. For ranked R-structures FP& = EXPR. 3.6. A bounded depth fixed point rule for DSOm
We define the bounded depth fixed point rule for digital second order logic as we defined the fixed point rule for this logic but with the following updating scheme.
Ao(t,8)=0
for all (21,8) E A' x P(A"') x . . . x P(A".)
co(;El,8)= 0
for all ( U , 8 ) E A' x ?(A"') x . . . x ? ( A a s )
AZ+l(t,U) =
P+l(G,U) =
HEi (Ti,8)if Ai(t, 8)= 0 and i 5 nL
Ai@, 8) otherwise F$ (21, 8)if A i ( E , 8)= 0
c(a,8) otherwise.
58
Notice that the only difference in this new sheme is that the number of steps in the inductive definition is forced to be at most ne where n = IAl and f? is a natural number.
Example 3.8. The function 7 defined in Example 3.7 actually uses n updates and therefore it is defined with only the bounded depth fixed point rule. Definition 3.18. Bounded depth fixed point digital second-order logic for R-structures, denoted B F P g , is obtained by augmenting DSOm with the maximization rule and the bounded depth fixed point rule. Remark 3.6. The number of updates is closely related to the running time of algorithms. That is why for EXPR we needed second-order structures to describe the entire computation. For PARR we only need polynomially many updates. On the other hand, for PARR exponentially many objects must be described -the gates of a circuit. Here second-order structures are used to name these objects. Theorem 3.6. B F P g = PARR. Remark 3.7. Theorems 3.4, 3.5 and 3.6 were shown for ranked Rstructures. It is possible to avoid this hypothesis and define the ranking within ~ D S O R .Roughly, one first considers a digital function R : A2 + ( 0 , l ) describing a total order in A. The existence of R is stated within ~ D S O R .Then, a number term r(u) for the ranking is defined as a fixedpoint for first-order logic. 4. Implicit Complexity over R
Implicit complexity has its roots in the work of S. Bellantoni and S. Cook, found in [3], with a viewpoint more related to the notion of recursion, and formal definition of functions, leading to a hierarchy of subsets of the set of primitive recursive functions.
4.1. Safe recursive functions Safe recursive functions are defined in a similar manner as primitive recursive functions. However, for safe recursive functions, we define two different types of arguments, each of which has different properties and different purposes. The first type of argument, called “normal” arguments, is similar to the arguments of our previously defined partial recursive and primitive
59
recursive functions, since it can be used to make basic computation steps or to control recursion. The second type of argument is called safe and can not be used to control recursion. This distinction between safe and normal arguments ensures that our safe recursive functions can be computed in polynomial time by a BSS machine. In our notations, the two different types of arguments are separated by a semicolon ";" : on the left part of the argument set we place the normal arguments and on the right part the safe arguments. We next define safe recursive functions. To do so, we consider functions: + R", taking as inputs arrays of elements in R", and returning as output an element in R". To simplify notation, elements in R" will be represented with overlined letters and elements in lR will be represented by letters. Also, u.3 will stand for the element in Roowhose first component is a and the remaining ones are 5. When the output of a function is undefined, we use the symbol 1.
Definition 4.1. We call basic functions the following three kinds of functions (which have only safe arguments): (i) functions making elementary manipulations of words of elements in R. For any a E R , 5 , % , 6E R": hd(;a.:)
=a
hd(;0) = 8 tl(;u.z) = z tl(; 0) = 0 cons(; u . E , G ) = u . 6 (ii) initial functions:
(iii) test function :
C ( ;a.z,5,a) =
y ifa=1
{-F otherwise
60
The set of safe recursive functions over R is the smallest set of functions + R", containing the basic safe functions, and closed under f : the following operations (1) Safe composition. Suppose g : (R")2 -+ R", hl : R" + R" and h2 : (R")' + R" are given functions. Then the composition of g with hl and h2 is the function f : (Rw)2-+ R":
m ; Y )= g ( h l ( q ;h2(55;g)) Note that it is possible to move an argument from a normal position to a safe position, whereas the reverse is forbidden. Suppose g : + Roois a given function. One can then define with safe composition a function f by f(Z, $7;
z)= g(z;y, z)
but a definition like the following is not valid:
f (qy, 5) = g(z,y;5). (2) Safe recursion. Suppose f 1 , . . . ,f k : (1~")~+ and 91,. . . ,gk : Q R " ) ~ + -+ ~ R" are given functions. Functions h l , . . . ,hk : + R" can then be defined by safe recursion:
Safe recursion captures polynomial time. Theorem 4.1. The set of safe recursive functions over R is the class FPR. 4.2. Safe recursion with substitution
Safe recursive functions with substitutions are a variant of safe recursion. They also have normal and safe arguments and rely on the same set of basic functions.
61
Definition 4.2. The set of substitution-safe recursive functions over R is the smallest set of functions f : (R")k + R", containing the basic safe functions, and closed under safe composition and the following operation:
Safe recursion with substitutions. Suppose f 1 , . . . , f k : (R")3 -+ R" and 91,. . . ,gk : (R")kee+3 + ]R" are given functions. Also, let g i j : R" -+ R" be given for i = l , . .. , k and j = l , . .. ,!. We call these aij substitution functions. The functions h l , . . . hk : (R03)3 -+ Rooare defined by safe recursion with substitutions as follows: hl(B,z';E,g), ... , h k ( O , z ; E , y ) = fl(z;E,y), ... ,frc(Z;E,g) and
if V i hi (Z,Z; y) #I
otherwise
if V i hi@,%%)#I
otherwise Safe recursion with substitution captures parallel polynomial time. Theorem 4.2. The set of substitution-safe recursive functions over R is the class FPARR. References 1. J.L. BalcBzar, J. Diaz, and J. Gabarr6. Structural Complexity I. EATCS Monographs on Theoretical Computer Science, 11. Springer-Verlag, 1988.
2. J.L. BalcBzar, J. Diaz, and J. Gabarr6. Structural Complexity II. EATCS Monographs on Theoretical Computer Science, 22. Springer-Verlag, 1990. 3. S. Bellantoni and S. Cook. A new recursion-theoretic characterization of the poly-time functions. Ciomputational Complexity, 297-110, 1992. 4. L. Blum, F. Cucker, M. Shub, and S. Smale. Complexcity and Real Computation. Springer-Verlag, 1998.
62 5. L. Blum, M. Shub, and S. Smale. On a theory of computation and complexity over the real numbers: NP-completeness, recursive functions and universal machines. Bulletin of the Amer. Math. Soc., 21:l-46, 1989. 6. 0. Bournez, F. Cucker, P. Jacob6 de Naurois, and J.-Y. Marion. Computability over an arbitrary structure. sequential and parallel polynomial time. In A. Gordon, editor, Proceedings of FOSSACS’03. Springer-Verlag, 2003. 7. F. Cucker. PR # NCR. Journal of Complexity, 8:230-238, 1992. 8. F. Cucker. On the complexity of quantifier elimination: the structural approach. The Computer Journal, 36:400-408, 1993. 9. F. Cucker and M. Matamala. On digital nondeterminism. Mathematical Systems Theory, 29:635-647, 1996. 10. F. Cucker and K. Meer. Logics which capture complexity classes over the reals. J. of Symb. Logic, 64:363-390, 1999. 11. H.-D. Ebbinghaus and Flum. Finite Model Theory. Springer-Verlag, 1995. 12. R. Fagin. Generalized first-order spectra and polynomial-time recognizable sets. S I A M - A M S PTOC., 7:43-73, 1974. 13. E. Gradel and Y.Gurevich. Metafinite model theory. In D. Leivant, editor, Logic and Computational Complexity, pages 313-366. Springer, 1996. 14. E. Gradel and K. Meer. Descriptive complexity theory over the real numbers. In J. Renegar, M. Shub, and S. Smale, editors, The Mathematics of Numerical Analysis, volume 32 of Lectures i n Applied Mathematics, pages 381-404. American Mathematical Society, 1996. 15. N. Immerman. Relational queries computable in polynomial time. Information and Control, 68:86-104, 1986. 16. K. Meer and C. Michaux. A survey on real structural complexity theory. Bulletin of the Belgian Math. Soc., 4:113-148, 1997. 17. C. Michaux. Une remarque ti propos des machines sur IR introduites par Blum, Shub et Smale. C. R. Acad. Sci. Paris, 309, Skrie I:435-437, 1989. 18. C.H. Papadimitriou. Computational Complexity. Addison-Wesley, 1994. 19. S. Smale. Mathematical problems for the next century. In V. Arnold, M. Atiyah, P. Lax, and B. Mazur, editors, Mathematics: Frontiers and Perspectives, pages 271-294. AMS, 2000. 20. M. Vardi. Complexity of relational query languages. In 14th annual A C M Symp. on the Theory of Computing, pages 137-146, 1982.
63
COMPUTABILITY, DEFINABILITY AND ALGEBRAIC STRUCTURES
ROD DOWNEY School of Mathematical and Computing Sciences Victoria University
P. 0. Box 600 Wellington,
New
Zealand
1. Introduction
In a later section, we will look at a result of Coles, Downey and Slaman [15Ia of pure computability theory. The result is that, for any set X , the set {deg(F)‘ : X is computably enumerable relative to F } always has a least member. While we now know that this theorem has an easy proof, it is nonetheless quite surprising and has consequences for the theory of Abelian groups. In these notes I hope to lead the reader into understanding this connection, and framing this result in the context of computability considerations yielding insight into classical algebra, and, in particular, coding in algebraic structures. In the first section, we will look at computability questions for algebraic structures, and in particular the sensitivity needed for the presentations of such structures. In the second section we will examine the question of assigning measures characterizing, at least to some extent, the algorithmic complexity of the relevant structures. The main tool here will be Turing aThis paper is an expanded version of an invited lecture “Every Set has a Least Jump Enumeration” given by the author during June, 1999 in Hsi-Tou, Taiwan, as part of the 7th Asian Logic Conference. This work is partially supported by the New Zealand Marsden Fund for Basic Science. I wish to thank the organizers of this conference for their hospitality and beautiful organization. Additional thanks to Reed Solomon, Richard Coles, Denis Hirschfeldt, and the referee, who supplied comment and corrections. Since there has been an extensive delay in the publication of this work, I have taken the liberty of addng some further references, especially with respect to the material on computably enumerable reals at the end.
64
degree, although many other classification tools are possible. In section 3 we will look at presentations of isomorphism types. In section 4 we will look at jump degrees of isomorphism types, reflecting possible codability. This leads into the problem for torsion free Abelian groups. In the 5th section, we will look at the least jump theorem, and its connection with enumeration degrees. In section 6 , we turn to analysis, where we will look at some material about presenting real numbers. Finally in the last section we offer some thoughts on what constitutes a good question on effective mathematics, and offer some questions we feel are in this category. The plan is to highlight a number of themes and open questions. While some of the material is familiar, much is new. I hope that the paper will cohere with, and complement, the material in my earlier survey [18], as well as the the long articles of Remmel 1911 and the author [19] and many other articles in the Handbook of Recursive Mathematics [34] and the book by Rosenstein [loo]. I assume that the reader is familiar with the rudiments of computability/recursion theory. Hopefully the reader will have met a priority argument, but certainly be familiar with the standard coding tricks that reduce computability theory to the study of partial functions on the N. (And reduce complexity theory to studying languages in E*.) (,) denotes a standard pairing function computably and bijectively mapping N x N e N. It is monotone in both variables. All uses etc, are bounded by s at stage s. Otherwise, notation is standard and follows, for instance, Soare [107].
2. Presentations The central classification tool in classical algebra such as group theory or linear algebra is isomorphism. Two structures are regarded as the same if they are isomorphic. This does not seem reasonable if one is concerned with trying to understand the algorithmic content of algebra. That is, if one views algebraic structures through computational eyes, and asks questions such as “if I am given a structure in such an online or computable manner so that I can perform the basic operations computably, then what else can I perform computably?” A good example is that of linear orderings. What is a reasonable definition of a computable linear ordering? Well, a linear ordering is a pair consisting of a set A and a binary relation 5 , which obeys the usual ordering axioms. So a reasonable definition for a computable linear ordering would be for the universe to be a computable set, usually identified with
65
N,upon which the order relation is a computable relation: given a, b E W,
we can decide which of a + b, b + a, or a = b holds. Now classically we have students think of an ordering of type w as being {0,1,2,3, ...}, with the usual order. But this is also isomorphic to an ordering listing 12 1 1000 lo5' + 2 + .... Any permutation of N gives rise to an ordering of type w in this way. So there are 2N0 differing presentations of orderings of type w. Here for a given structure A, a presentation will be some structure isomorphic to A. The degree of such a presentation B is simply the degree of its open diagram. In the case of a linear ordering of N, this is simply the degree of the ordering relation. It is obvious that many of the 2N0 presentations of w will have differing computability properties since there are only countably many Turing machines. However, not completely obvious is the fact that differing computable presentations can have differing computability properties. For instance, the adjacency relation A(x, y) on an ordering A is defined to hold on a pair x,y if x 4 y and for all z , it is not the case that x + z 4 y. The following is an easy example, taken from the folklore.
+ +
+
Observation 2.1. There is a computable ordering (A,+) of type w for which the adjacency relation is not computable. Proof. We begin with A0 = (25 : x E N} ordered in the usual way. We devote the pair 4e,4e 2 to the satisfaction of the e-th requirement Re below.
+
Re : pe(.,.) is not an adjacency function for A. Here qe denotes the e-th binary partial computable function. It will have output 0 or 1 with 0 meant to represent the fact that the points represented by the arguments are nonadjacent, and 1 adjacent. We will add at most one point in the middle of the currently adjacent pair 4e, 4e 2. To this end we set aside the numbers {(e,s) : s E N} for the sake of Re. The action is that, while Re is not declared satisfied, if we see a stage s > e where ( P ~ , ~ ( y) X , $= 1, we declare Re as satisfied and put 4(e, s) 1between 4e and 4e + 2. This clearly makes sure that pe cannot be the adjacency function for A, since either we never need to act for it because it never tells us that 4e, 4e+Z is an adjacency, or it does so tell us, and we make it wrong by adding a point between them. Since we declare it satisfied, we do this at most once. Hence each point has at most finitely many predecessors. Finally the ordering A so constructed is Computable, since we decide the fate of even points at the beginning and to decide if 4x 1 is in A (and
+
+
+
66
where), we simply compute the pair e, s such that ( e ,s) = x . Then if s 5 e , 42 + 1 $! A , otherwise 42 + 1 is in A iff we put it in between 4e and 4e 2 at stage s, and no other. 0 A few comments on the “wait and see”proof above. First it admits many variations. For instance, we could define as, say, Cenzer and Remmel [12], an ordering to be polynomial t i m e if the universe is (0,l)’ (there are other variations) and the relation x 4 y is polynomial time in the length of x and y . The argument above can easily be modified to show that there is a polynomial time ordering where the adjacency relation is not computable! Additionally, notice that the argument demonstrates that we can have computable orderings which are isomorphic yet not computably isomorphic (since the computability of the adjacency relation would be preserved by a computable isomorphism). This is not always the case. Define a computable ordering to be computably categorical if every computable ordering isomorphic t o it is computably isomorphic to it.
+
Theorem 2.1. (Remmel [go], Goncharov [44]) A computable linear ordering is computably categorical iff it has a finite number of adjacencies. A direct proof of the Remmel-Goncharov theorem works by observing that the usual Cantor back and forth categoricity of the rationals can be effectivized for the “if” direction, and by a finite injury priority argument based on guessing and killing adjacencies for the hard direction. We remark that the argument is as easy as the one above in the case that we have a computable list of adjacencies. However this is not always the case.
Theorem 2.2. (Downey and Moses [30]) There i s a computable linear ordering A such that for any computable linear ordering B isomorphic t o A , the adjacency relation o n B has Turing degree 0‘, the degree of the halting problem.
So we have seen that for some order types there can be many computable presentations, and for some only one. One might ask what else is needed t o specify a structure up t o computable isomorphism. In the case of w adjacencies are enough. Theorem 2.3. (essentially a special case of Moses [83]) Suppose that two linear orderings of type w each have computable adjacency relations. T h e n the orderings are computably isomorphic. Proof The proof is fairly easy. Suppose that A and B are the relevant orderings. Without loss of generality, let ao, bo be the least members of each.
67
Using complete search, and the computability of the adjacency relation, we can find the successor of a0 in A and the successor of bo in B , and then map each t o the other, etc. 0 Ash and Knight (e.g. [l,691) have formulated various extensions of the conditions of Theorem 2.3 which give decidability conditions guaranteeing the computable categoricity of classes of computable well orderings or discrete orderings. As one would expect, the conditions connect the finite iterations of the “condensation” operator (two points become identified if they are a finite distance apart) with infinitary decidability conditions. Similar considerations arise in proof theory in connection with uniqueness of paths through Kleene’s 0. Turning to general presentability questions, already we can formulate two open questions.
Question 2.1. Classify, by order type, the linear orderings that have computable presentations, or show that this is impossible. Question 2.2. Does there exist a linear ordering that has presentations of each nonzero Turing degree, but no computable presentation? We finish this section with some comments about Questions 2.1 and 2.2. Regarding 2.2, Wehner [115] and Slaman [lo21 have each constructed examples of relational structures which have presentations in each degree except zero. Hirschfeldt, Khoussainov, Shore and Slinko [52] have found several canonical reductions allowing results on presentability to be transferred to other structures such as groups, rings, integral domains and the like. Thus (for example) they can construct examples of groups that have presentations in each nonzero degree but are not computably presentable. This reflects earlier work on Bore1 reducibility by, for instance, Friedman and Stanley [41] (see also Kechris [54].) The Hirschfeldt et. al. methods don’t apply t o orderings. One notable result here is due to MilIer.
Theorem 2.4. (Miller [Sl]) There is a linear ordering with presentations in each A: degree except 0. Harizanov suggests that one can define the spectrum of a structure A as the degrees in which A has presentations. Similarly for a class of degrees C, Miller defined the C-spectrum of A to consist of the intersection of the spectrum of A with C. So the Theorem above can be rephrased as : there is a linear ordering whose A: spectrum is ever A; degree except 0 .
68
Actually this is related to earlier work on the degree spectrum of a relation R on a model d. The spectrum of R is the possible degrees of R in various computable presentations of A. The notion was first articulated by Harizanov, who constructed models with, for instance, two element degree spectra. Notice that if we consider the adjacency relation on a linear ordering, then by the Downey-Moses result, we get the following.
Corollary 2.1. There is a computable ordering A such that the degree spectrum of the adjacency relation has a single element 0‘. Actually, the degree spectrum of the adjacency relation can be fairly rich. Theorem 2.5. (Downey) There is a computable linear ordering A f o r which the degree spectrum of the adjacency relation consists of every nonzero computably enumerable degree. Proof This proof uses a technique of Downey and Knight [26]. Let Q be a linear ordering. Define y(Q) = (q 2 q) . Q. That is, we replace each point of Q by a copy of the rationals (q) followed by a single adjacency followed by the rationals. Notice that the collection of adjacencies of y(Q) is isomorphic to Q. Downey and Knight [26] proved that Q has a presentation of degree a’ iff y(Q) has a presentation of degree a. In particular, if Q is chosen to be the linear ordering of Miller’s theoremb which has presentations Qb of each computably enumerable degree b, then certainly y(Q) is a computable linear ordering with no computable presentations with computable adjacency relation. (To see this suppose that is a computable presentation of y(Q) with computable adjacency relation. From this we can construct a computably enumerable subset of with the same ordering as isomorphic to Q. This is easily seen to be isomorphic to a computable ordering, a contradiction.) To finish the proof we need to look into the guts of the Downey-Knight proof. Let Q be given and suppose that Q is a computably enumerable linear ordering of degree b. Via Cantor’s proof, we can consider Q as a l-Iy subset of the rationals with the standard ordering. We will build r(Q) as a computable ordering so that the the degree of the adjacency relation in r(Q)is the same as the degree of Q. We can computably embed the standard copy of Q into a computable copy of (Q 2 Q) . Q, where we
+ +
6
6
6
6
+ +
bStrictly speaking, we need the fact that if a linear ordering has a presentation of c.e. degree a then it has a c.e. presentation of degree a. This is not difficult to prove.
69
regard Q as being a subset of the last copy of Q. Thus we can regard the x-th rational as being identified with a canonical adjacency in y(Q). Note that if Q = n,Qs is a computable approximation to Q, then x E Q iff
VS(X E Q s ) . The idea is simple. We do nothing for x unless we see a stage s such that x @ Q,. At such a stage s we put a copy of Q in the successivity representing x. (Strictly speaking we would have an infinite list of fresh numbers to represent rationals we add, and we would be doing everything with finite subsets, but this obscures the idea.) Clearly y(Q) so constructed is computable, and the set of adjacencies of this copy of y(Q) has, in fact, 0 the same m-degree as Q. This gives the result, An interesting question is to look at the possible spectra a linear ordering can have. We know, by a result of Knight [63], that if L has a presentation of degree a then it has one of each degree above a. Another example of a structure closely related to orderings is that of boolean algebras. While it is possible to prove things directly, it is nice to extract results for boolean algebras from results for linear orderings via the following technical device.
Theorem 2.6. (The Representation Theorem, folklore after Stone) Every countable boolean algebra is isomorphic to an interval algebra Intalg(L) of a linear ordering of the same degree. (Recall that Intalg(L) is the algebra generated by the finite unions of left closed right open intervals of the ordering L.) Proof. This is just an effective version of Stone’s theorem. We represent a given B as an algebra of sets. Thus let B = U , B , with Bo = (0, l}, and B,+l - B, = { b , } . We will define at each stage a subalgebra B, containing B, . Define Lo to be the ordering with two points labeled 0 and 1. (Then IntaZg(L0) consists of the two sets 0 and [0, l).) So we have the induced mapping go with 0 I+ 0 and 1 I+ [O,l). At stage s 1 we will have a set of Atoms, = { a s l ,...,a,,,} listing the atoms of the subalgebra of B, together with the linear ordering L, = O,zsl,...,x,, = 1 so that g s ( a s j ) I+ [ x , ~ -xsj) ~ , inducing an isomorphism from the subalgebra B, to Intalg(L,). we need to do nothing. Otherwise, for each At stage s 1 if b, is in a = aSi such that b, splits a (i.e. both a A b, and a A b, are nontrivial), add a new point y to Ls+l between x,j-l and xSj to split the interval [xsj-l,xsj)into [ X , ~ - ~ , YU) [y,xSj).Naturally we map asj A b, to one of them, say, [ x , ~ -y~),and map asj A to [ y ,xSj). Note that this generates h
+
h
h
+
70
-
two new atoms for B,+1, and of Intalg(L,+I). Let c1, ...,c, denote the atoms of B,+1 below b,. Clearly we have ensured that the induced map g(b,) = g(c1) U ... U g(cm) works. The results follows. Notice that in the proof of the Representation Theorem we get a 1 - 1 correspondence between atoms in B and adjacencies in L. We might well expect that we could have similar results concerning the relation A ( z )=“z is an atom” and its spectrum, as we had for the adjacency relation. However, the situation is rather different. We need the following Lemma.
Theorem 2.7. (RemmeEVaught [91])“ (i) Suppose that B is any boolean subalgebra of Q with infinitely many atoms {di : i E N}. Let 6 be the subalgebra of Q obtained from B b y splitting each atom of B a finite number of times in Q (together with B). (Thus the atoms of B^ would be { e i l , ..., ein(.) : i E N where di = ei, V ... V einci,}.)Then B is isomorphic to B. (ii) (Rephrasing (i) in terms of linear orderings). Suppose that L and 2 are linear orderings with infinitely many adjacencies, and g : L 2 is an order-preserving embedding such that (a) if [z, y ] is finite in L then [ g ( z ) , g ( y ) ]is finite in 2,and (b) if z E 2 is not in the image of L, then there are x,y in L such that [zC,ylis finite and z E [ g ( z ) , g ( y ) l . Then Intalg(L) E Intalg(2). A
Remmel used Theorem 2.7 to prove many of his results on boolean algebras. For instance, to prove that if a boolean algebra has infinitely many atoms then it is not computably categorical, we need to guess atoms and then split them if the range looks like an atom. So long as the splitting is finite, then we don’t leave the isomorphism type. An easier example is to construct a boolean algebra isomorphic to the given one whose atom set has no infinite computably enumerable subset. We refer the reader to Remmel [91] for details. We have seen that the adjacency relation for a linear ordering can be intrinsically complete, since its spectrum can consist of a single degree 0’. This is never the case for the atomicity relation of a boolean algebra.
Theorem 2.8. (Downey [17])Every computable boolean algebra is iso‘In his Thesis, Vaught proved this result for atomic boolean algebras and Remmel extended the result to all countable boolean algebras.
71
morphic t o one where the collection of atoms is Turing incomplete. This theorem is proven using Theorem 2.7 and a fairly complicated infinite injury priority argument. To demonstrate the technique, we sketch the proof of a much easier result. By the work of Jockusch and Soare [60], we know that there are linear orderings L with presentations of low degree (i.e. L’ --T 0’)not isomorphic to a computable linear ordering. This is not true of boolean algebras.
Theorem 2.9. (Downey and Jockusch [25]) Suppose that B is a low boolean algebra. T h e n B is isomorphic t o a computable boolean algebra.
Proof. Using the proof of Stone’s theorem above we can regard B = I n t a l ( L ) with L low, and notice that the adjacency relation for L is computable from 0’since L is low. Hence, by the Limit Lemma we can suppose we have a computable approximation L ( x ) = lim, L,(x) where the fact that a pair is an adjacency or not will be known from some point on. Without loss of generality, we suppose that L has endpoints xo, X I . We map xi I+yi. Now a new point appears between xo and XI. This new point a may or may not be in L , but since L has infinitely many points there is no apparent harm in putting a new point b between y1 and yo, and matching. If the point a leaves, then we can find the next least point (by Godel number) a‘ and rematch a‘ e b. The only problem is if we have some interval ( a l , a ~ ) currently matched to ( b l , bz). A new point c appears between a1 and a2 and we respond with a new point d between the bi. The point c leaves, but since our order must be computable, d cannot leave. If the ai I+ bi are good mappings then we don’t want to move them. Here is where we get to use lowness. We know that if (a1,az) looks often enough like an adjacency it is one. So we only respond to the appearance of new points between the ui if it appears that ( a l , a z ) is not an adjacency. Either it really is not, in which case we are safe with new elements between the bi, or after a while we will constantly believe that ( a l , az) is an adjacency, and hence we will have mapped the interval ( a l , az) to ( b l , bz) where ( b l , bz) contains a finite number of points not in the range of our mapping. Now we can apply the 0 Remmel-Vaught Theorem to get the isomorphism. The theorems above suggest the following open questions. Question 2.3. What can be said about the possible spectra for the relation <‘uis an atom” for a boolean algebra B? Question 2.4. What can be said about the possible degrees of boolean
72
algebras not isomorphic to computable ones? We remark that, with respect to 2.3, Goncharov modified the coding argument of Feiner [35] to construct a computable boolean algebra whose atoms are intrinsically non-low, for any n E N. (See Remmel[91].) Thurber [113] extended the Downey-Knight theorem above to show that every low2 boolean algebra is isomorphic to a computable one, and this has been extended to each low4 boolean algebra by Knight and Stob [57]. New ideas will be needed to show that, as the author conjectures, every low, boolean algebra is isomorphic to a computable one. More generally it is very interesting to wonder as to the possible degree spectra of relations on models. There are many possible versions here. Two are (i) the general spectra problem where we look at the possible degrees in all models, and (ii) the degrees of the relation in computable models. The current best result for two-element degree spectra for computable models is provided by the following result of Hirschfeldt. Theorem 2.10. (Hirschfeldt [51]) Let a be any nonzero computable enumerable degree. Then there is a computable structure A and a relation R on A whose spectrum is ( 0 , a}. Amazingly, we don’t even know if there is some upper bound on the second degree a in a two-element spectrum containing 0 . For instance it is completely consistent with our current state of knowledge that there could be a computable structure and a relation with two element spectrum 0 , a such that a is not even arithmetical! We remark that Hirschfeldt has more generally proven the following regarding two element spectra. Theorem 2.11. (Hirschfeldt [51]) Let b,a be any two computable enumerable degrees. Then there is a computable structure A and a relation R on A whose spectrum is { b, a}. Theorem 2.11 even works for w-c.e. degrees. We remark that AshCholak-Knight [3] have shown that, in some sense, the A: degrees are the important ones. They almost show that if a relation hits all A! degrees then it then it contains all degrees. More precisely, they show the following. ~
dWe note that because the set of realizations of a given relation is a C: set, there are partial orderings of c.e. degrees that cannot be degree spectra. Furthermore, we also know that if the degree spectrum contains a nonhyperarithmetical degree and a hyperarithmetical degree, i t must be uncountable.
73
Theorem 2.12. (Ash, Cholak, and Knight [3]) For a relation U on a computable structure A, the following are equivalent. (i) For every A! degree a there is a computable structure B and an isomorphism f : A I3 such that both f and f ( U ) have degree a. (ii) For every degree b there is a structure B and an isomorphism g : A e B such that both f and f ( U ) have degree b.
*
In his thesis, Hirschfeldt has proven that the inclusion off in the degree constraints is necessary, since he has constructed, for each n, a computable structure and a relation whose spectrum is exactly the A: degrees. For more on these issues we refer the reader to Hirschfeldt’s thesis, KhoussainovShore [57], or [34]. We remark in passing that these issues are closely related to the algorithmic dimension of the structures. The algorithmic dimension is the number of computable isomorphism types of computable models. We have seen that structures such as the rationals can be computably categorical, and hence have dimension 1, and can be like w which has dimension 00. Goncharov [43] constructed a computable structure of dimension n for each n with 1 5 n 5 00. It is also known that if a structure has a computable copy not computably isomorphic to it but A: isomorphic to it, then the structure has dimension 00. Therefore if the dimension is 2, the isomorphism between the models must be non-A;. All known examples are A!. The following question is opene.
Question 2.5. (Khoussainov) Are there isomorphic computable structures A, B of dimension 2, such that the isomorphism between them is not A!? How complicated can it be? Can it be not arithmetical? We remark that these issues are related to definability in the structures. For instance, a relation on a structure A is called, following Ash and Nerode [6], intrinsically (e.g.) computable, if it is computable in each computable copy of A. Essentially, if the model has enough decidability, then a relation is intrinsically computable iff it is described by a computable L,,,, formula. And similarly, a structure is computably categorical iff it is described by a computable Scott family, provided that it obeys certain decidability conditions. (The reader is directed to the excellent survey [57] for a thorough discussion of these and related matters.) =Recently, Downey, Goncharov, and Hirschfeldt have found a technique which seems to solve this positively.
74
We now turn to Question 2.1, which asks us to classify, by order type, the linear orderings that have computable presentations, or show that this is impossible. The real question is, what is an acceptable classification of the order types? Computability theory can provide an excellent vehicle for demonstrating that no reasonable classification exists, by showing that the relevant class is not Borel. Two nice examples are the following. Slaman and Woodin examined the following question. Given a partial ordering (P,5 ) ,a linear extension is a linear ordering (P,5 ) on the same universe P such that 2 5 y implies x 5 y. It is not difficult (Rosenstein [loo]) to prove that each computable partial ordering has a computable linear extension. It is even true that each dense computable partial ordering has a dense computable linear extension. We remark that it is also classically known that a scattered partial ordering has a scattered linear extension. (Recall that an ordering is called scattered if it does not embed the rationals.) It is unknown if this result effectivises, a question of Rosenstein.
Theorem 2.13. (Slaman and Woodin [103]) {e : We is a computable partial ordering of N without a dense linear extension } is a Cf complete set. This answers a question of Lo6 by showing that there is no reasonable classification of the countable partial orderings with dense linear extensions. A similar example is provided by the theory of difference sets. Recall that a set of integers A is called a difference set iff there is a set B of integers such that A = D(B), where
D ( B ) = (2 - y : y < 2 E B } . Downey, F’uredi, Jockusch and Rubel [21] first examined inverting the difference operator effectively. They showed, for instance, that there are computable sets that are the difference set of a co-c.e. set but not the difference set of any computable set. Many other results are proven in [21], including one that has a classical spinoff every subset of N is the intersection of two difference sets. The underlying motivating question was due to Ruzsa who, in the ~ O ’ S , asked for a reasonable classification of the sets that are difference sets. In particular, Rubel reformulated the question to ask if the set was Borel. Again computability theory provides a negative answer.
Theorem 2.14. (Schmerl [lll])The set of difference sets is C t complete.
75
Schmerl’s paper contains many other good things besides, and is beautifully written.
3. Degrees of Isomorphism Types The natural guess for definition which assigns a degree to an isomorphism type would be the least degree of any presentation in that type. For instance, the degree of dense linear orderings without endpoints is clearly 0 since we can computably represent Q. Linda Jean Richter was the first author to systematically study the degrees of structures. (Richter [94])It turns out that some types have degrees and some do not. One example of a structure with a degree is a finitely presented group. If G = ( 2 1 , ...,z n l y l , ..., y m ) is finitely presented then its degree is simply the degree of its word problem. Note that if H is isomorphic to G then we only need to look a t the images of {XI, ..., z n } t o obtain generators of H and through this, H can compute the open diagram of G. Below are some other examples of structures with nontrivial degrees. We obtain these examples using a preliminary theorem of Richter. This theorem encapsulates the so-called “combination method.”
Theorem 3.1. (Richter [94,95])Suppose (i) and (ii) below hold for a degree a and a theory T over a finite language L. Then there is a structure of L whose isomorphism type has degree a. (i) There is an infinite computable sequence of finite structures {Ai : i E N} such that Ai i s not embeddable into Aj f o r i # j . (ii) For each S C w , there is a structure As such that: (iia) As is a countable structure of T . (iib) As ST S . (aic) Ai is embeddable into As iff i E S . Proof. Fix a degree a and let D be any set in a. Consider ADeB. Evidently AD,^ ST D. Now suppose that B E A D e ~ .Then i E D iff A2i is embeddable into B , and i E iff Azi+l is embeddable into B . It follows 0 that D ST B and so the isomorphism type of A D e has ~ degree a. Theorem 3.1 has many applications. We list a couple below.
Theorem 3.2. (Richter [94,95]) Let a be any degree. (i) There is a n Abelian group whose isomorphism type has degree a. (ii) There is a lattice whose isomorphism type has degree a. (iii) There is a graph whose isomorphism type has degree a. Proof. They are all basically similar. We do (i). Consider Ai = Z p i , the cyclic group of order p i , where pi denotes the i-th prime. The structure As
76
0 is the standard direct sum of the Ai. Now we apply Theorem 3.1. At this stage we should point out that, as with many of these results on structures and their degrees, Theorem 3.2 has several consequences in the degrees. In particular Theorem 3.2 was used by Richter t o demonstrate that If 0"' ST a and @ is a j u m p preserving automorphism of the degrees then !@(a) = a. On the other hand, some structures do not have degrees. For instance we have the following.
Theorem 3.3. (Richter [94]) If a n order type has a degree, that degree is 0 .
Proof Sketch To prove Richter's Theorem, Theorem 3.3, we will actually sketch a proof of the following. Given a linear ordering L then there is an ordering 2isomorphic to L such that the infimum of the degrees of L and 2 is 0 ; that is, they form a minimal pair. Basically we use the finite extension method. So for requirement e , we diagonalize against ae,rewhich say that if =:?I = f , total, then f i s computable. We will have already specified a finite portion of 2,and a finite partial isomorphism ge from an initial segment of L (say, up to and including e ) to an initial segment of 2. Then we ask if there is any way to extend L, so as t o cause a disagreement. (The only proviso here is that the extension must be consistent with the order type. This causes a little bit of concern, and we refer the reader to Richter's paper or Downey [18] for more details.) If we can, then we cause a disagreement. If not then we can argue that the common value is computable. 0 We remark that the proof of Stone's theorem in the previous section also shows that isomorphism types of boolean algebras don't have nontrivial degrees. h
Corollary 3.1. (Richter 1941) If the isomorphism type of a Boolean algebra has a degree t h e n that degree is 0. The proof follows by Richter's proof of the linear ordering case. As above, she showed that for any noncomputable linear ordering L there is an ordering 2 isomorphic to L such that the infimum of the degrees of L and 2 is 0 . 4. Jump Degrees and a Question on Torsion Free Groups
Since orderings and boolean algebras don't have isomorphism types with degrees, Jockusch suggested that this might be because of the number of
77
quantifiers needed to code in such structures. With this in mind he suggested the following definition.
Definition 4.1. (Jockusch) Let A be any structure. Then the n-th jump degree of A is min{deg(D(n)) : B
A}.
The idea is that the measure is coarser than the notion of degree. For linear orderings, we don't seem to get much for n = 1.
Theorem 4.1. (Knight [63]) If a h e a r ordering has a first jump degree it is 0'. Knight's theorem is proven by a forcing argument. However, once we get to n = 2 all manner of coding is possible.
Theorem 4.2. Let n 2 2 be any computable ordinal. Let a > O(").Then there are linear orderings L and J such that (i) (Ash, Jockusch, Knight [4])L has n-th jump degree a and no n - 1-st jump degree. (ii) (Downey-Knight [26]) J has n-th jump degree O(") and no n- 1-st jump degree.
In the context above, Downey and Jockusch, see [18], began a study of jump degrees of torsion free Abelian groups. Recall that G = (G, +) is called torsion free if no element has finite order. Such groups are classically of interest since they are, for instance, precisely the Abelian groups that can be ordered. They are of interest computability theoretically as we now see. We remark that in Abelian groups with torsion we can use the Ulm sequence to code a great deal of information. This leads to a well behaved degree and jump degree theory analyzed in Lin [75,76], Smith [112], Goncharov [44],and, for jump degrees, Oates [SS]. However, the structure theory and the computability theory become more subtle if we assume that the group is torsion free. Khisamiev has studied these objects, and has obtained many fundamental results. Torsion free Abelian groups behave very interestingly. For instance, Feiner [35,36] proved that there are computably enumerably presented linear orderings and boolean algebras not isomorphic to computable ones. This is not true for torsion free Abelian groups, although it is true for Abelian groups in general.
78
Theorem 4.3. (Khisamiev [63]) If G = (G,+,=) i s a C1 presented torsion-free group t h e n G is isomorphic t o a computably presented group. In the classical language of combinatorial group theory, every recursively presentable torsion free Abelian group is isomorphic to one with a solvable word problem. Khisamiev’s Theorem, Theorem 4.3, solves affirmatively a question of Baumslag, Dyer and Miller [7]. This question arose from the study of the integral homology of finitely presented groups, where computably enumerably presented torsion free Abelian groups turn out to be the natural objects of study. We also remark that for linear orderings if L is 1: presented, (that is, the ordering relation is II:) then L is isomorphic to a computably presented ordering. On the other hand, as we see below, there are II? presented torsion free Abelian groups not isomorphic to computable ones. This was originally an observation of Khisamiev [63] after Khisamiev and Khisamiev [64]). We give a much simpler proof than the complex Khisamiev and Khisamiev one, and show that G can be chosen to have rank 1. (The rank of a group is defined below.) We then explore the notion of jump degree for rank 1 groups. We remind the reader of some definitions, etc. from Fuchs [42]. Every countable (X-presentable) torsion free Abelian group is isomorphic to an additive (X-presented) subgroup of V,, where V, denotes the vector space over Q with basis {ei : i E N} (ei is the vector that is everywhere zero except for 1 in position i). If G is isomorphic to a subgroup of then the least such n is called the rank of the group. The simplest and best understood of the torsion free Abelian groups are those of rank 1. These are effectively isomorphic to subgroups of (Q, +, =). Consequently we really need only consider subgroups of (Q, =). Let p l , p z , . . . be the primes in increasing order. The pheight h p ( a )of a E G, a subgroup of Q, is defined via
w,
+,
hp(a)= k if k is greatest with p k I a (in G) and hp(a)= 00 if p k I a for all k. For a E G we define the characteristic a , x ( a ) via
79
It is not difficult t o see that as G has rank 1, if a, b E G with a, b # 0 then x ( a ) and ~ ( bare ) equivalent in the sense that they differ only finitely often. We write x ( a ) =* ~ ( b )We . can thus define the type of x ( G ) to be the =* equivalence class of x ( a ) for any a # 0 in G. The classical theorem of Baer is Theorem 4.4. (Baer, see e.g. Fuchs [42])] If G and H have rank 1 then G H iff G and H have the same type. Proof. The proof of Theorem 4.4 is easy. In the nontrivial direction take g E G and h E H with x ( g ) =* ~ ( h )From . this it follows that ma: = n g is solvable in G iff m y = nh is solvable in H . The isomorphism is given by x t-+ y . 0 We remark that, since for all a: E G, such m, n exist, the isomorphism
given by the proof above is effective relative to the presentations. We also note that this means torsion free Abelian groups of rank 1 are computably categorical in the sense that G E H implies G Z c H , where Sc denotes computable isomorphism. The same is in fact true of finite rank torsion free Abelian groups: The work of (e.g.) Goncharov [44],Ash-Nerode [6] or direct constructions show the following. Theorem 4.5. A computable torsion free Abelian group is computably categorical iff it has finite rank.
+, =).
Let G be a subgroup of (Q,
Define the standard type of G S(G)
via
S ( G ) = { ( i , j ) : j 5 the i-th member of x ( G ) } for some fixed a E G. Note that if G is X-presented, then S(G) is E f . Let A be Cf. Let A^ = {(x,O),( y , 1) : a: E N , y E A } . Note that A zm2. Then it is easy to construct an X-presentable G with S ( G ) = A^. (In fact given any finite type sequence T = (ao,a1 ,...), with ai < 00 for all i, which is given by a function which is EF there is an X-presentable group G ( T ) with this type sequence.) It is now easy to see
Theorem 4.6. There is a rank 1 II:+l-presented group G n o t isomorphic t o a II:-presented group.
torsion free Abelian
Proof. We do the case n = 0 and then relativize. Choose A to be C!j!-complete and construct a @‘-presentable G with S ( G ) = 2. Then by our earlier work we can suppose G is II1 and as S ( G ) is Cq - C y , by Baer’s theorem, G cannot be computably presentable lest S(G) be Ey.
80
We remark that Khisamiev and Khisamiev [64] construct a 11: computable torsion free Abelian group F with a computable basis and a maximal divisible subgroup N of F with FIN in II: - A:. Their construction is much more intricate. We also remark that there is related work by Fkiedman, Simpson and Smith [39]. What about the degrees of isomorphism types? Using a fairly simple argument, Julia Knight [unpubl.] showed that there exist torsion free Abelian groups of arbitrary degree. Downey [18] improved this to rank one groups.
Theorem 4.7. (Downey, after Knight) Let a be any degree. Then there exists a torsion free Abelian group G of rank one with degree a. Proof. Let S be any set of degree a. Note that if A is any set with S @ sE Ef then S @ s
=
Theorem 4.8. (Downey and Jockusch) (i) There is a rank 1 G (of finite type) with proper l-degree O'f (ii) More generally, f o r every degree a 2 0' there is a torsion free Abelian group of proper l-degree a. (iii) Similarly, f o r all degrees b 2 0" there is a torsion free Abelian group of rank 1 with proper %degree b.
For a proof, see Downey [MI. The situation becomes very interesting for jump degrees and rank one groups. This is a beautiful example of classical mathematics providing stimulus to classical computability theory. We say that a group G has finite type if for all i, IN(2) n S(G)I < 00. So for instance, the group of Theorem 4.7 has finite type. That is, G has no elements of infinite height. The following is the key nexus between Baer's theorem and jump degrees. fHere an n-degree, or rather n-th jump degree, is called proper, and there is no n - l-st jump degree.
81
Observation 4.1. (Downey and Jockusch) Every torsion free Abelian group of finite type and of rank 1 has 1-degree iff the following question has a positive solution. 0
Question. Is it true that for all sets M , the collection of degrees {deg(B') : M is computably enumerable in B } has a least element?
A proof of Observation 4.1 can be found in full detail in Coles, Downey, and Jockusch [15]. It is not difficult. There is a similar version saying that every torsion free Abelian group of rank 1 has 2-jump degree iff for all sets X, the set of degrees below has a least member. {deg(B)" : X is Cf} Downey and Jockusch conjectured that the question above has a negative solution. Furthermore they spent much effort trying to find counterexamples. As we will see in the next section, there is a reason for their failure to find such a counterexample.
5. Every Set Has a Least Jump Enumeration The above leads us to consider the sets C(A) = {X I A E CF}, and, c(A) = {deg(X) I A E Cf}. (That is, the sets that A can be enumerated from, and their degrees.) More generally, we have seen that there is some interest in the sets cJI")(A)= min{deg(X)(n) I A E C,"}. The question of Downey and Jockusch was open for several years, until it was found that it has a positive solution. This is a very surprising result. The original proof was a complicated priority argument. This was later simplified to the forcing argument from Coles-Downey-Slaman [15]. (The following argument differs from the final published one which uses the technique of forcing the jump. We present this alternative proof because it is interesting from the point of view of the material on enumeration degrees to follow.)
Theorem 5.1. (Coles, Downey, Slaman [15]) For all X, cl(X) exists.
82
Proof. Given a set A we want to construct a set F such that A is CF and F' ST X' for all sets X with A E Cf . Definition 5.1. Let X C N. 0
0
0
The partial order Pf is the set of finite enumerations of subsets of X ordered by extension. So an element of PF is a function, p say, from {0,1,. . . , k} into X, for some Ic E w. For p and q in Pf,we say p extends q in Pf and write p >f q , if the graph of p is contained in the graph of q. We use 1 t o denote the empty function, and of course 1 E Pf.
We fix a Pf generic enumeration of A. ClearIy A is Cfl since y E A if and only if there is an z such that (2,y) E 61. Let F be the forcing relation for IIY(61) sentences. That is,
F = (03,'p) I P @
Jeor
('p
E
m G 1 ) and P
v CP)}.
We first observe that A _
X' for all X E Cf
Proof For any p and
'p,
( p , 'p) E F if and only if either
(1) P # p f , or (2) there is a proper extension q of p in
Pf such that q Iy 9.
Since Pf LT A , in the first case we have that p E Pf is computable in X ' for any set X for which A is C f . In the second case, we have an existential condition that refers to a finite amount of positive information about A. This condition is Cf for sets X for which A is C f , and hence, for such an X , case 2 is computable in X'. Therefore F ST X' for any X such that A is E f . 0 Lemma 5.2. There is a set G such that A is CY(G) and
Proof. We construct G sense of ~ f .
G'
F so that every fact about G' is forced in the
83
Step 0: Let po = 1. Step s 0
0
+ 1 = 2e : If (p,, cp) E F, then let p,+l be the least extension of p , in P;' such that p,+l It ~ c p . Otherwise, (p,, cp) 61 F. Let p,+l = p , since p , It cp.
+
Step s + 1 = 2e 1 : Let x be the least number in A not in the range of p,. Let p,+l be the least extension of p , with x in its range. At the end of the construction define G = UsEw p,. Now clearly A is Cf. Observe that for case 1 of the even stages of the construction, we can find p,+l computably in A. Hence A @ F is an oracle that can decide every IIy condition about G. Therefore G' S T A @ F. We have already seen that A ST F and hence, G'
Theorem 5.2. (Coles, Downey, Slaman [15]) For all n >_ 1, and for all X , cjl"'(x> exists. We get the corollaries we are after.
Corollary 5.1. (Coles, Downey, Slaman [15]) Suppose that G is a rank 1 torsion free Abelian group. Then G has second jump degree. Furthermore, if G has finite type, G has first jump degree. Subsequently, Soskov (private communication) found an elegant proof of the above based on the theory of enumeration degrees. (Actually, if we look closely, the proofs are essentially the same.) Recall that A S e B iff there is a computably enumerable set W such that for all x ,
x E A iff 3u[(x,u)E W A D, C B ] . Here D, denotes the u-th canonical finite set. It is known that enumeration reducibility can be strongly connected with algebra. For instance, Charles Miller I11 (see [50]) proved that given two finitely generated groups GI and Ga, G2 can be embedded into a group finitely presented relative to GI iff G2 is enumeration reducible to G I . This is
84
sometimes called the relative Higmann Embedding Theorem. Enumeration reducibility is also the key here. There is a canonical embedding of the semi-lattice of the Turing degrees into the semi-lattice of the enumeration degrees. The Turing degree of a set A is identified with the enumeration degree of the set A @ A, d, ( A@ A). For any set A define A+ = A @ A. A set A is called total if A -,A+. An enumeration degree is total if it contains a total set. So the Turing degrees are the total enumeration degrees. Notice that A is computably enumerable in B iff A 5, B+ and A
C ( A )= {X : X is total and A 5, X } Lemma 5.3. (Soskov) C ( A ) has a n element of least degree ifld,(A) is total. Proof. Clearly if & ( A ) is total, then & ( A ) is the desired least degree. Assume that there exists X O E C ( A ) ,such that X O ieX for all X E C ( A ) . Then we have for all total X ,
A5eX
*XO
<e
X.
Therefore, by Selman’s Theorem (see [108]), X O A and hence A -,X O . 0 The enumeration jump was defined by Cooper and further studied by K. McEvoy. (See, for example, [77].) It is defined as follows. Given a set A, define
where Q z denotes the z-th enumeration operator. Set the enumeration jump of A to be the set A’, = (K,f)+. McEvoy [77] studies the basic properties of the enumeration jump. Notice that A’, is a total set and for total sets the e-jump is enumeration equivalent to the Turing jump.
Lemma 5.4. (Soskov) For every set A, &(A’,) is the least among the degrees of the elements of
C‘(A)= {X‘: X is total and A 5, X } . Proof. By the monotonicity of the e-jump operation, A: 5 X‘ for all X E C ( A ) . It remains to show that there exists a total X such that A 5, X and A: X‘. This follows directly from [108], Theorem 1.2. 0
=,
a5
Soskov also provided a proof of the case for more general n > 1. Define by means of induction the set Consider n sets Ao, , . . ,A,-1. P(A0,.. . , An-l): 1) P(Ao) = Ao; 2 ) P ( A 0 , .. . ,Ak+l) = P(A0,.. . ,Ak); CB
Ak+l. Let
C'(Ao,.. . ,An-l) = {X@): A. is Cl(X), . . . ,An-l is C , ( X ) } . Theorem 5.3. (Soskov) d,(P(Ao,. . . ,A,-l);) is the least among the degrees of the elements of the set C'(Ao,. . . ,A,-I). Proof. Let P = P(Ao,. . . , An-l). It is sufficient to show that there exists a total X such that X ( " ) zeP,' and Ai is Ci+l(X). Take a total Q such that P 5, Q and P,' G Q'. By [108], Theorem 1.2, there exists a total X such that X ( " - l ) =e Q and Ai is C & , i = 0,. . . , n- 1. Clearly X ( " ) E e P,'. Take A. = . . . = An-2 = 0 and A,-l = A . Then P ( A 0 , .. . ,An-l) = @(,-') @ A . Hence d , ( ( @ ~ ( ~ @ - l )A ) ; ) is the least among the degrees of the 0 elements of C(")( A ) . Soskov observed that his techniques solved a couple of the questions listed at the end of a preliminary version of [15]. One left open was question 2 there, which he interpreted to mean the following. Question 5.1. Characterize the sets X for which the enumeration jump is equivalent to the Turing jump. 6. Computably Enumerable Reals The results at the end of the previous section demonstrated how reducibilities other than Turing reducibility can yield significant insight into the effective content of classical mathematics. Another example of this phenomenon is the use of Ziegler reducibility, a variation of quasi-reducibility (see Downey, LaForte, Nies [29]), to characterize relative algebraic closure in groups. (See [50].) Yet another example is provided by the degrees of computably enumerable bases for subspaces of V, , characterized precisely as the weak truth table degrees below that of the given space. (See Downey and Remmel [31]) In this section we will see yet another example of this phenomenon, and additionally see how this methodology applies to classical analysis. We begin with a simple question, asked for centuries, what is n real? There are several definitionsg, and we list several below. (We consider reals gThese definitions define the relevant real up to equivalence.
86
in (0,l). And no real is rational, for ease of presentation. We remark that an alternative approach is to only look at Cantor Space 2* with basic clopen sets [o] = {aa : a E 2*} which is not homeomorphic to (0,l) but measure-theoretically isomorphic. Here, of course, we are using Lebesgue measure, and have p ( [ a ] )= 2-14.) 0 0
0
a has an infinite (e.g.) decimal expansion, a = .alaz.... a = z n ~ A 2 - ~where , A c N. a = CnE~2-lnl,where A c C*. a is the limit of an increasing sequence of rationals.
Now a computable real would presumably be a computable limit of a computable sequence of rationals, using the Dedekind cut definition. More precisely, there is a computable sequence qi, i = 1,2, ... of rationals such that for any n we can compute an m such that la-q,J < 2-n. Now it is not difficult t o show that this definition coincides with a having a computable decimal expansion, or a = CncA2-n for a computable subset A of N. (See Rice [92].) However, we would get a little stuck if we also claimed that this ( some computable A (0, l}*,as we now is the same as a = C n E ~ 2 - l nfor see. Define a computably enumerable real to be the limit of a computable sequence of rationals. That is, we have a = lim,q, such that we can compute qi for all i. But note that although we know that the sequence converges, we don’t computably know how fast, at least in general. The following theorem is from Calude, Hertling, Khoussainov and Wang [lo]. A set A is called prefix-free iff for all (T E A, and all T with (T an initial segment of 7, r # A.
Theorem 6.1. (Calude et. al. [lo]) The following are equivalent.
(i) a is cornputably enumerable. (ii) The set L ( a ) = {q E Q : q < a } is computably enumerable. hPrefix-free sets are considered for technical reasons since if a set A is prefix-free then, by Kraft’s inequality, we know that C n ~ ~ 2 - l nconverges, l and hence A is measureable. Prefix-free sets are crucial in a proper treatment of (e.g.) randomness, and other complexity issues, for c.e. reals, a fact first realized by Chaitin [ll]. Under Chaitin’s observation of the effectiveness of Kraft’s inequality, shows that computably enumerable prefix-free sets are the measures of the domains of partial computable prefix free machines, in the same way that computably enumerable sets are the domains of partial computable functions.
87
(iii) There is an infinite computably enumerable prefix free set A with a = Cn~A2-ln1. (iv) There is a computable prefix free set A such that a = &A2-lnl. (iv) There is a computable function f ( 2 ,y ) of two variables, such that (iva) If, for some k , s we have f ( k , s + 1) = 0 yet f ( k , s ) = 1 then there is some k' < lc such that f (k',s) = 0 and f (k',s+ 1) = 1. (ivb) a = 'alaz... is a dyadic expansion of a with ai = lim, f ( i , s ) .
(v) There is a computable increasing sequence of rationals with limit ai. The reader should note that (iii) implies (iv) because we always have many strings we can potentially add at some stage. The reader should also note that although things coincide for computable reals, there are computably enumerable reals b which are not, what we will call, strongly computably enumerable.
Definition 6.1. A real a is called strongly computably enumerable if there exists a computably enumerable B g N such that nEB
The fact that there are c.e. reals that are not strongly c.e. was first noted by Soare [106]. Here is a very quick proof. To defeat W e ,the e-th c.e. set, we make our real have 1 in the 4e-th place, and 0 in places around it. We need do nothing unless 4e enters We. At that stage, we can make the 4e'th place of our real become 0 by making the 4e - 1-st place become 1. The motivating question for most of the rest of the section is the following. How can we present our c.e. real? Part of the answer is provided by Calude, Coles, Hertling and Khoussainov [9]. Those authors examined the cut definition of real, and its effective content. They defined a representation A of a (c.e.) real a as an increasing sequence of rationals qi for i E A with limit a. They asked what types of degrees can A have. Already we have seen that A can be computable. It is not difficult to prove the following.
Lemma 6.1. (Soare [106], Calude et. al. [9]) Let L ( a ) denote the lower 'If we only ask for a limit of a computable sequence of computable reals, then instead of getting the computably enumerable reals we gets real whose dyadic expansion is computable from the halting problem. ([53])
88
cut of a real a. That is, L ( a ) is the collection of rataonals less than a. Then if A is a representation of a, A
0=
c
2-14,
M(0T)J.
where M is a universal prefix-free machine. This quantity is Chaitin’s famous Omega-number, the so-called halting probability. We study c.e. reals under various reducibilities which preserve randomness or segments, such as modified forms of m- and a modified form of weak truth table reducibility. For instance, in [28], we looked at Soloway reducibility. It would take us a little far to motivate and examine this material in detail, but as an example, for strongly c.e. reals Solovay reducibility becomes the following. We say a = &&42-n sw-reduces to b = CnEB2-n iff there is a constant c and a weak truth table procedure @ with use cp, such that 0 0
@ B = A and for all n, cp(n)5 c . n.
The idea is that if I am close to b I can compute a rational close to a. The general case is similar but operates on cutsj. In an unpublished manuscript, Solovay proposed the reducibility above, and demonstrated that it preserves the “degree of randomness.” That is if a 5 s p then there is a constant d such that, for all n, the Kolmogorov complexity of a 1 n is less than or equal t o that of p r n plus d. It is interesting to examine the degree structure. For the c.e. reals, the top degree is the Solovay degree of the halting problem, [0]s.R is a set that is naturally wtt-complete but not tt-complete. It is also illuminating that again the relevant reducibility jThat is, for reals a and p we say that a is Solovay reducible to p iff there is a partial computable function f with domain in Q, and a constant c E M such that for all all ~ f(q)l. rationals q < P, f(q) J. and IP - q ) 2 2 ‘ 1 ~-
89
turns out to be stronger than T-reducibility. We refer the reader to Calude et. al. [lo, 28,9] for more details. Additional work has happened in the last few years such as the paper of Downey, Hirschfeldt, and Nies in the present volume. Solovay's work and much new paterial will appear in the forthcoming book of Downey and Hirschfeldt [22]. Returning to our story, we note that Calude et. al. proved the following very interesting generalization of Lemma 6.1.
Lemma 6.2. (Calude et. al. [9]) Suppose that A represents a. Then A is an infinite half of a splitting of L ( a ) . Proof. Clearly, if A represents a then A must.be an infinite c.e. subset of L(a). The thing to note is that L(a) - A is also c.e.. Given rational q , if q occurs in L(a), we need only wait till either q occurs in A or some rational bigger than q does. 0 Note that this means that if a is computable then every representation of a is computable. Also note that the proof actually gives that if A represents a, A L(a). The following is easy to show using the technique of the proof above.
Lemma 6.3. (Calude et. al. [9]) Let A be a representation of a . For subsets B of A , the following are equivalent.
B represents a. B is half of a splitting of A . It would be nice to be able to reverse the above. That is, prove that every splitting of L ( a ) is a representation of a. Clearly this is false since some splittings won't even be representations. Calude et. al. [9] did however prove that (i) a has a representation of degree deg(L(a)), and (ii) every representation can be extended to one of degree deg(L(a)). Here is their proof. Let qi e a , be a given c.e. sequence of rationals converging to a. We build the splitting of L(a) and the representation of a of degree deg(L(a)) in stages. At stage 0, let TO = qo. Suppose that at stage s we have T O , ...,T',(,) with qi = rn(i) for each i 5 s. At stage s let g8 be the largest Godel number seen so far. Consider the rationals q with qs < q 5 qs+l and such that the Godel number of q is less than or equal to the maximum of gs and the Godel number of qs+l. Suppose that there are m such q . Then we let n ( s 1) = n ( s ) m, and let r,(,)+k be those q in increasing order for k = 1, ..., m.
+
+
90
The sequence ri is increasing and converges to a. Hence {ri} l W t t L(a). Conversely, to decide if q E L ( a ) ,compute a stage s where g, exceeds the Godel number of q. Assuming that q is not yet in L ( a ) ,then since q > rn(,), q E L ( a ) iff q E {ri}. Note that this is in fact an m-reduction. Actually, with a little care, we can improve the Calude et. al. [9] result, and get a complete characterization of the representations of a in terms of the weak truth table degrees, and later m-degrees, of splittings of L ( a ) . Theorem 6.2. (Downey) The following are equivalent:
b is the weak truth table degree of a splitting of L ( a ) . b is the weak truth table degree of a representation of a. Proof. To prove Theorem 6.2, we need only show that if L ( a ) = C U D is any c.e. splitting of L ( a ) then there is a representation C = {ci} of a of wtt degree that of C. (Without loss of generality, we suppose that C is noncomputable.) We do this in stages. At each stage s, we assume that we have enumerated C, and D, so that L ( a ) , = C, U D,, where L ( a ) , is the collection of rationals in L ( a ) by stage s, including all those of Godel number 5 s. Additionally we will have a parameter m(s). At stage s 1 compute C,+l and Ds+l. Find the least rational, q E Cs+l,by Godel number, if any, such that q > m(s). If no such q exists, set m(s 1) = m(s),and do nothing else. If one exists, put q into C,+l and reset m(s 1) to be the maximum rational in L(a),+l. To verify the construction, first note that is an increasing sequence of rationals. Its limit will be a provided that it is infinite, because of the use of m(s). First we claim that m ( s ) -+ 00. Suppose not, so that there is an s such that, for all t 2 s, m(s)= m(t).Then we claim that C is computable, this being a contradiction. To decide if z E C , go first to stage s' = s g ( z ) , where g ( z ) denotes the Godel number of z. If z # C,, , then either z > m ( s ) , or z E D,). In either case, z # C. Hence m(s)+ 00. Note that Swtt C because only numbers entering C enter 2 and can do so only at the same stage. Finally we see that C Lwtt To decide if q E C , find the least stage s such that s > g ( q ) and E, 1 g ( q ) 1 = 1 g(q) 1. If q 5 m(s)then as above we can decide if q E C. Otherwise, q > m(s). Suppose that q # C3+1.We claim that q # C. Otherwise, consider the stage s1 where q enters C. Now, if q > m(s1- 1)then either q or some even h
+
h
+
+
e
+
e.
+
+
91
e.
smaller number must enter Thus it can only be that q 5 m ( s 1 - 1). As a consequence, there is some least stage t where s 5 t < s1 where m(t)< q and m(t 1) 2 q. Consider any stage t‘ 2 s where m(t‘) # m(t’ 1 ) . For an induction, we suppose that q # Ctf. We only reset m(t’)because we saw some q‘ enter Ctt U Dtt with q‘ > m(t’) 2 m(s). We put q’ into Since g ( q ) 1 = c^ g(q) 1, we can only have that the Godel number of q’ exceeds q 1. Therefore, it cannot be that q enters C at stage t‘. If it did, then either q or some number with Godel number less than or equal to q would enter in place of q’. (Remember here, we are considering in C, U D, all rationals in the associated lower cut with Godel number 5 u.) Thus, in particular, at stage t = t’, q # Ct. But if q , or some number with Godel number below q does not enter 6 at stage t + l , which it cannot, then q is not in C. Hence C Swtt 2. 0 The proof above can be improved to give the following.
+
e8r
+
+
+
r
+
e
Theorem 6.3. (Downey) a is the m-degree of a c.e. splitting of L ( a ) iff a is the m-degree of a representation of a . Proof. The modification is the following. At stage s, when we increase m(s),we put into C not just the least q but, in rational increasing order, all q entering C with Godel number less than s. Now as we have seen, since m(s) + 00, C is a representation of a. Moreover, the same argument shows that C E w t t We claim the reductions are m-reductions. First IrnC. Given q go to a stage s bigger than the Godel number of q. If q is below m ( s ) then, as before, we can decide computably if q E C. Else, note that q E C iff q E The same argument shows that C 6.0 We remark that many of the theorems of Calude et. al. [9] now come out as corollaries to the characterization above, and known results on splittings and wtt degrees. We refer the reader to, for instance, Downey and Stob [32]. For instance, we get the following. h
-
e.
e.
srn
Corollary 6.1. There exist computably enumerable reals ai such that the collection of T-degrees of representations R ( a i ) have the following properties. (a) R ( a 1 ) consists of every c.e. ( m - ) degree (ii) R ( a 2 ) f o r m s a n atomless boolean algebra, which is nowhere dense in the c.e. degrees. (iii) There is a c.e. degree b with 0 < b < deg(L(a3)) such that if x is the degree of a representation above b then x = deg(L(as)), and i f
92
x is below b then x = 0.
0 Proof. See Downey and Stob [32]. We also remark that the above has a number of other consequences regarding known limits to splittings. For instance; Corollary 6.2. If a c.e. real a has representations in each T-degree below that of L ( a ) then either L ( a ) is Turing complete or low. Proof See Downey [20]. 0 As our final topic we look at the other form of representing reals. To avoid confusion we have the following definition. Definition 6.2. Let A
c {O,l}*.
(i) We say that A is a presentation of a c.e. real a if A is a prefix free c.e. set with a =Cn~A2-l~'.
(ii) We say that a c.e. set A is a strong presentation of a if a = &A
(Here, the
. xu.
.xu denotes the binary expansion.)
Previously we have seen that a has representations of degree L(a). However, presentations can behave quite differently. Theorem 6.4. There is a c.e. real a which is not computable, but such that if A presents a then A is computable.
Proof Sketch We briefly sketch the proof, details being found in Downey and LaForte [28]. We only mention the case (i) above. We must meet the requirements below.
Re : We presents a implies We computable. We build a presentation of A , via the nearly c.e. definition. That is, we have an approximation a = .ao,s... and obey the conditions that if ~ i = , 1 and ~ ai,,+l = 0 then aj,,+l becomes 1 for some j < i. To make a noncomputable, we must also meet the requirements:
Pe : For some i , i E We iff ai = 0. The strategy for Pe is simple. We must pick some i to follow it, and initially make it 1. At some stage s, if we see i enter W e ,then we must make ai,t = 0 for some t 2 s.
93
To make this cohere with the Re we need a little work. First, we need to surround i with some 0’s so that we can make it 0 by making the one before, say, 1. However, more importantly, we need to also make sure that for those Rk of higher priority if wk presents a then wk is computable. Associated with RI, will be a current “length of agreement”. This is a number m ( k , s ) such that a , - CnEwb,,2-lnl< 2 - ” ( I , y S ) . We can assume that a, 2 CnEwk,, 2-lnl, for otherwise, we can win by withholding numbers from a. We promise that once m ( k , s ) > d, then no number of length 5 d can enter Wk. Now the idea is that when we see some P, require attention for e bigger than k , if i is smaller than m ( k , s ) , the interesting case, then we wish to put a relatively big number into a , by changing position i for the sake of P,, yet we wish to not allow numbers of low length to enter wk. The idea is to slowly work backwards. So first we will make position into A,+1. m ( k ,s) + 1 = 1 by adding something of length We then do nothing until WI,responds by giving us a stage t > s with at - ~ ~ ~ ~ ~ < , 2-m(k?s)+1. ~ 2 - l n l Note that wk can only change on strings of long length, since we only changed A slightly. Now we repeat, adding another string of the same length 2-m(k*s)+1into As+l. Again we wait for another expansion stage. Note that this next addition changes things at position m(k,s) or earlier. We can continue in this way at most 2m(k+)-i many times till we get to change position i. Now there are two outcomes. Either at some stage, we don’t get recovery, so that WI,does not present A , or wk responds at each stage and we get a change only on long strings. This means that we can compute wk. This is the basic module for a standard IIz priority argument, akin to the embedding of 1-3-1. Details can be found in Downey and LaForte [28]. 2-m(k1s)+1
0
We remark that the proof above is very similar to the embedding of 1-3-1 in the c.e. Turing degrees. The result is related to lattice embeddings since Downey and LaForte did show the following. Theorem 6.5. (Downey and LaForte [ 2 8 ] )I’L(a) has promptly simple degree, then a has a noncomputable presentation.
We remark that all of this is related to wtt degrees. As our final results we offer the following.
94
Theorem 6.6. (Downey and LaForte [28]) Suppose that A presents a A. Then there is a presentation of a of wtt degree c.e. real a. Let B
B. Proof Sketch Suppose that I?* = B is a wtt reduction with use y and A is a presentation of a. We suppose that every stage is expansionary. We build a presentation C Ewtt B. We suppose that 0 # B. For each i 5 r ( n ) ,and each o with [oil= i, choose a length m 2 n, and strings rk(Ui)of length m for k 5 p(ai,n)large enough, so that the set of r ' s can be chosen to keep C prefix free, and such that there are 2m - 2%many T k (oi) of lengthk m. At stage s assume that there is a unique element n in B,+1 - B,. For each i 5 $n), Ioil = i, if oi E A,+1 - A,, put coding markers rk(oi,n)of length m = m(ui,n) into Cs+l, causing i to enter C,EC,+12-1T1. If j > y(n) and oj of length j enters A,+I, put 2-('J>0) into C. It is not difficult to argue that the set C works. n) with First C Swtt B. To see this, to decide if /3 enters C compute (7, /3 = ( r , n ) . If n = 0 compute a stage s where B 171 = B, 1 171. Then /3 E C iff /3 E Cs+l. If n # 0 then since /3 can only enter at C the stage n enters B we get C Swtt B . Conversely, to see that B SWttC , to decide if n enters B simply compute a stage s where C, 1 Tk(Oi,n)= C 1 Tk(oi,n)for all i 5 y ( n ) ,and k 5 p(ai7 n). It is clear that C presents the real since when i enters the A-sum, it is because some oi enters A. At such a stage, we will put enough I - ( ( T ~ , n ) into 0 C to cause i to enter the C-sum. 2m("'7")-i
r
Corollary 6.3. Suppose that a is strongly c.e. with a = ' X A . Then B is the wtt degree of a presentation of a iff B swttA. Proof Put Oi-'l E B iff i E A. This is a presentation with m-degree that of A . Now apply the theorem. 0 Downey and Terwijn [33] proved a major extension of the above. Since {lo : (T E A} U {Or : T E B } has the same wtt- degree of the join of those of A and B , counting quantifiers, by Theorem 6.6, we see that if a is a c.e. real than the wtt-degrees of presentations of a forms a C: ideal in the c.e. w t t degrees. kThe idea here is that we can use the C-sum.
T'S
collectively to add something of length i to the
95
Theorem 6.7. (Downey and Terwijn [33]) Suppose that Z is any X! ideal in the c.e. wtt-degrees. Then there is a c.e. noncomputable real a such that the wtt-degees of presenatations of a are precisely the members of 1. The proof of this theorem combines the “drip-feed” strategy of DowneyLaforte, a coding strategy, and a 2: approximation strategy. We refer the reader to [33] for further details. 7. Epilog: What are good questions in computable mathematics? Some of the first questions asked in effective mathematics must be those of Hilbert and of Dehn. Focusing on Dehn, he posed the famous word, conjugacy, and isomorphism problems for finitely presented groups. These questions gave rise to combinatorial group theory, and are primarily of interest because they give enormous insight into the structure of groups. This is the key. Good questions should give insight into either computability (as in our torsion free result) or need considerable algebraic or analytic insight to solve. We offer with some trepidation a couple below which we feel will fall into this category.
Question 7.1. A structure is called &-decidable if one can decide all nquantifier statements. For each n, is there a finitely presented group which is n but not n + 1 decidable? Is there a finitely presented group which is not decidable, but is n decidable for each n? Question 7.2. (Downey and Kurtz) For each II; class C is there a compuably presented torsion free group with the orderings (up to reversal) in 1-1correspondence with the members of C? Reed Solomon has observed that the answer is no if the group must be Abelian. Incidentally, Solomon [104,105] proved that a computably presented group is isomorphic to a quotient of a computably ordered version of a free group by a computable convex normal subgroup. This theorem has a remarkably difficult proof, going through a classical group ring construction and needing small cancellation theory. It would be interesting to know why this so, and whether this must be the case. Perhaps proof theory can provide the answer. Note that there is no known simple proof that the two generator free group can be ordered.
96
Question 7.3. (Friedman, Simpson, Smith [39]') Given a I I: C is there a commutative ring with 1 whose set of prime ideals is in 1-1correspondence with the member of C? Question 7.4. For each n, classify, by order type, the computable linear orderings with a nontrivial I: automorphism. The answer is known for n = 0 (Schwartz), but open for n = 1. See Downey [18] for more on this question. Question 7.5. (Rosenstein) It is known that a scattered linear partial ordering has a scattered linear extension. Is this computably true? What is its proof theoretical strength? Since the original writing of this paper, this question has be analyzed by Downey, Hirschfeldt, Lempp, and Solomon [24]. The proof theoretical strength is surprisingly high (around AT&.)
'Strictly speaking, Friedman et. al. claimed this as a theorem in [39]. It was only later that a flaw was found in the proof. The reverse mathematics result was proven by separating sets. They proved this and stated the problem in the later addendum [40].
97
References 1. Ash, C. J., Categoricity in the hyperarithmetical degrees, Ann. Pure and Appl. Logic, vol. 34 (1987), 1-34. 2. Ash, C. J., Labelling systems and r.e. structures, Ann. Pure and Appl. Logic, vol. 47 (1990), 99-120. 3. Ash, C., P. Cholak, and J. Knight, Permitting, forcing, and copying of a given recursive relation, Annals of Pure and Applied Logic, Vol. 87 (1997), 219-236. 4. Ash, C. J., C. Jockusch and J. F. Knight, Jumps of orderings, Trans. Amer. Math. SOC.Vol. 319, (1990), 573-599. 5. Ash, C. J., J. F. Knight, M. Manasse, and T. Slaman, Generic copies of countable structures, Ann. Pure and Appl. Logic, vol. 42 (1989), 195-206. 6. Ash, C. J. and A. Nerode, Intrinsically recursive relations, in Crossley [16], 26-41. 7. Baumslag, G., E. Dyer, and C. Miller 111, On the integral homology of finitely presented groups, Topology, Vol. 22 (1983), 27-46. 8. Boone, W. W., Certain simple unsolvable problems in group theory, IVI Nederlakad. Wentenschappen, Proc. Ser. A, vol. 57 (1954), 231-237, 492-497, V O ~58. (1955), 252-256, 571-577, V O ~60. (1957), 22-27, 227-232. 9. Calude, C., R. Coles, P. Hertling, and B. Khoussainov, Degree-theoretical aspects of computably enumerable reals, t o appear. 10. Calude, C., P. Hertling, B. Khoussainov, and Y . Wang, Recursively enumerable reals and Chaitin’s R number, in STACS’98, Springer Lecture Notes in Computer Science, Vol. 1373, 1998, 596-606. 11. Chaitin, G. A theory of program size formally identical to information theory, J. Assoc. Comput. Mach., Vol. 13 (1966), 547-569. 12. Cenzer, D. and J. B. Remmel, Polynomial time vs recursive models, Ann. Pure and Applied Logic, Vol. 54 (1991), 17-58. 13. Chisholm J., and M. Moses, Undecidable linear orderings that are nrecursive for each n, in preparation. 14. Coles, R., R. Downey, and B. Khoussainov, Initial segments of computable linear orderings. Order, Vol. 14 (1997-1998), 107-124 15. Coles, R., R. Downey, and T. Slaman, Every set has a minimal jump enumeration, Journal London Math. SOC.(2) Vol. 62 (2000), 641-649. 16. Crossley, J. N., (ed) Aspects of Effective Algebra, (Ed) Upside Down A Book Co., Yarra Glen, Vic. Australia (1981). 17. Downey, R., Every recursive boolean algebra is isomorphic t o one with incomplete atoms, Ann. Pure and Appl. Logic, Vol. 60 (1990), 193-206. 18. Downey, R. On presentations of algebraic structures, in Complexity, Logic and Recursion Theory, (A. Sorbi, ed.), Marcel Dekker, Lecture Notes in Pure and Applied Mathematics, Vol. 197 (1997), 157-206. 19. Downey, R., Computability theory and linear orderings, Handbook of Recursive Mathematics, Val 2, (Ed. Ershov, Goncharov, Nerode, Remmel, Marek), Studies in Logic Vol. 139, North Holland, 1998, 823-976. 20. Downey, R. G . , On the universal splitting property, Mathematical Logic
98
Quarterly, 43 (1997) 311-320. 21. Downey, R., Z. Feuredi, C. Jockusch, and L. Rubel, Difference sets and computability theory, Annals Pure and Applied Logic. Vol. 93 (1998), 6372. 22. Downey, R. and D. Hirschfeldt, Algorithmic Randomness and Complexity, Springer-Verlag monographs in Computer Science, t o appear. 23. Downey, R., D. Hirschfeldt, A. Nies, and F. Stephan, Trivial reals, this volume. (extended abstract in Computability and Complexity in Analysis Malaga, (Electronic Notes in Theoretical Computer Science, and proceedings, edited by Brattka, Schroder, and Weihrauch, FernUniversitat, 2946/2002, 37-55),July, 2002.) 24. Downey, R. G., D. Hirschfeldt, S. Lempp, and R. Solomon, Computabilitytheoretical and proof-theoretical aspects of partial and linear orderings, t o appear, Israel J. Math. 25. Downey, R. and C. Jockusch, Every low boolean algebra is isomorphic to a recursive one, Proc. Amer. Math. SOC.,Vol. 122, (1994), 871-880. 26. Downey, R. G. and J. F. Knight, Orderings with a-th jump degree O ( a ) , Proc. Amer. Math. SOC.,vol. 14 (1992), 545-552. 27. Downey, R. G. and S. A. Kurtz, Recursion theory and ordered groups, Ann. Pure and Appl. Logic, vol. 52 (1986), 137-451. 28. R. Downey and G. LaForte, Presentations of computably enumerable reals, Theoretical Computer Science Vol. 284 (2002), 539-555. 29. Downey, R. G., G. LaForte, and A. Nies, Enumerable sets and quasireducibility, Annals of Pure and Applied Logic, Vol. 95 (1998), 1-35. 30. Downey, R. G. and M. F. Moses, Recursive linear orderings with incomplete successivities, Trans. Amer. Math. SOC.,vol. 320 (1991), 653-668. 31. Downey, R. G. and J. B. Remmel, Classification of degree classes associated with r.e. subspaces, Ann. Pure and Appl. Logic, 42 (1989) 105-125 32. Downey, R., and M. Stob, Splitting theorems in recursion theory, Annals Pure and Applied Logic, 65 (1)((1993) 1-106). 33. , Downey, R. and S. Terwijn, Computably Enumerable Reals and Uniformly Presentable Ideals, Archive for Mathematical Logic Vol. 48 (2002), 29-40. 34. Ershov, Y . , S. Goncharov, A. Nerode, and J. Remmel, (eds) Handbook of Recursive Mathematics, (V. Marek, associate editor), Studies in Logic Vol. 139, North Holland, 1998. 35. Feiner, L. J., Orderings and Boolean Algebras not isomorphic to recursive ones, Thesis, MIT (1967). 36. Feiner, L. J., Hierarchies of boolean algebras, J. Symb. Logic, vol. 35 (1970), 365-373. 37. Feiner, L. J., The strong homogeneity conjecture, J. Symb. Logic, vol. 35 (1970), 373-377. 38. Feiner, L. J., Degrees of non-recursive presentability, Proc. Amer. Math. SOC.,V O ~ .38 (1973), 621-624. 39. Friedman, H., S. Simpson and R. Smith, Countable algebra and set existence axioms, Ann. Pure and Appl. Logic, vol. 25 (1983), 141-181.
99
40. Friedman, H., S. Simpson and R. Smith, Addendum to “ Countable algebra and set existence axioms,” Ann. Pure and Appl. Logic, vol. 28 (1985), 319320. 41. Friedman, H. and L. Stanley, A Bore1 reducibility for countable models, J. Symb. Logic, Vol. 54. (1989), 894-914. 42. Fuchs, L., Infinite Abelian GTOUPS,Vol. 11, Academic Press, 1973. 43. Goncharov, S. S., On the number of nonautoequivalent constructivizations, Algebra and Logic, Vol. 16 (1977), 169-185. 44. Goncharov, S. S., Autostability of models and Abelian groups, Algebra and Logic, Vol. 19 (1980), 13-27. 45. Grigoreiff, S., Every recursive linear ordering has a copy in DTIMESPACE (n,log(n)), J. Symb. Logic, vol. 55 (1990), 260-276. 46. Harizanov, V., Some effects of Ash-Nerode and other decidability conditions on degree spectra, Ann. Pure Appl. Logic. vol. 54 (1991), 51-65. 47. Harizanov, V., Uncountable degree spectra, Ann. Pure Appl. Logic. vol. 54 (1991), 255-263. 48. Harizanov, V., Turing degree of the non-zero member in a two element degree spectrum, Ann. Pure and Appl. Logic, vol. 60 (1993), 1-30. 49. Higman, G., Subgroups of finitely presented groups, Proc. Royal SOC.London, Vol. 262 (1961), 455-475. 50. Higman, G. and E. Scott, Existentially Closed Groups, Oxford University Press, 1988. 51. Hirschfeldt, D., Degree Spectra of Relations on Computable Structures, PhD Thesis, Cornell University, 1999. 52. Hirschfeldt, D., B. Khoussainov, R. Shore and A. Slinko, Degree spectra and computable dimension in algebraic structures, in preparation. 53. Ho, Chun-Kuen, Relatively recursive reals and real functions, Theoretical Computer Science, Vol. 219 (1999), 99-120. 54. Kechris, A. S., New directions in descriptive set theory, Bull. Symbolic Logic, vol. 5 (1999), 161-174. 55. Khoussainov, B. and A. Nerode, On automata representable structures, in preparation. 56. Khoussainov, B. and A. Nerode, Automatic Model Theory, monograph in preparation. 57. Khoussainov, B. and R. Shore, Effective model theory: the number of models and their complexity, to appear. 58. Hird, G., Recursive properties of relations on models, Ph.D. Thesis, Monash University (1983). 59. Jockusch, C. G., Ramsey’s theorem and recursion theory, J. Symb. Logic, V O ~ .37 (1972), 268-279. 60. Jockusch, C. G. and R. I. Soare, Degrees of orderings not isomorphic to recursive linear orderings, Ann. Pure and Appl. Logic., Vol. 52 (1991), 39-64. 61. Jockusch, C. and R. Soare, Boolean algebras, Stone spaces, and the iterated Turing jump, Journal of Symbolic Logic, Vol. 59 (1994), 1121-1138. 62. Khisamiev, N., Connections between constructivizability and nonconstruc-
100
63. 64.
65. 66. 67. 68. 69. 70. 71. 72. 73. 74. 75. 76. 77. 78. 79.
80. 81. 82. 83. 84. 85.
tivizability for different classes of Abelian groups, Algebra and Logic, Vol. 23 (1984), 220-233. Khisamiev, N., Hierarchies of torsion free Abelian groups, Algebra and Logic, Vol. 25 (1986), 128-142. Khisamiev, N., and Z. Khisamiev, Nonconstructavizability of the reduced part of a strongly constructive torsion-free Abelian group, Algebra and Logic, Vol. 24 (1985), 69-76. Kierstead, H., Recursive ordered sets, in I. Rival (ed) Combinatorics and Ordered Sets, Contemporary Math., vol 57, Amer. Math. SOC.(1986). Kierstead, H., G. McNulty and W. Trotter, Recursive dimension for partially ordered sets, Order, vol. 1, (1984), 67-82. Knight, J. F., Effective constructions of models, in Logic Colloquium (ed. Paris, Wilkie and Wilmers) North-Holland, Amsterdam (1986). Knight, J. F., Degrees coded into jumps of orderings, J. Symb. Logic, vol. 51 (1986), 1034-1042. Knight, J. F., A metatheorem for finitely many workers, J. Symbolic Logic, V O ~ .55 (199), 787-804. Knight, J. F., Constructions for transfinitely many workers, Ann. Pure and Appl. Logic, vol. 48 (1990), 237-259. Kopytov, A., and V. Kokorin, Fully Ordered Groups, John Wiley and Sons, 1974. Lerman, M., On recursive linear orderings, in Lerman, Schmerl and Soare [1981], 132-142. Lerman, M., Degrees of Unsolvability, Springer-Verlag, New York (1983). Lerman, M. and J. Rosenstein, Recursive linear orderings, in Metakides [1982], 123-136. Lin, C., Recursion theory on countable Abelian groups, Ph. D. Thesis, Cornell University, 1977. Lin, C., The effective content of Ulm’s theorem, in Crossley [16], 147-160. McEvoy, K., Jumps of quasi-minimal enumeration degrees, J. Symb. Logic, Vol. 50 (1985), 839-848. McNulty, G., Infinite ordered sets, a recursive perspective, in I. Rival (ed) Proc. of Symp. on Ordered Sets, D. Reidel, Dortrecht (1982). Metakides, G. and A. Nerode, Recursion theory and algebra, in Algebra and Logic (ed. J. N. Crossley), Lecture notes in Math., vol. 450, New York (1975), 209-219. Metakides, G. and A. Nerode, Effective content of field theory, Ann. Math. Logic, vol. 17 (1979), 289-320. Miller, R., The A! spectrum of a linear ordering, t o appear. Moses, M. F., Recursive Properties of Isomorphism Types, Thesis, Monash University, Melb. Australia (1983). Moses, M. F., Recursive linear orders with recursive successivities, Ann. Pure and Appl. Logic, vol. 27 (1984), 253-264. Moses, M. F., Relations intrinsically recursive in linear orderings, Z. Math. Logik Grundlagen. Math., vol. 32 (5), 467-472. Moses, M. F., Decidable discrete linear orderings, J. Symb. Logic, vol. 53
101
(1988), 531-539. 86. Oates, S., Jump Degrees of Groups, Ph. D. Thesis, University of Notre Dame, 1989. 87. Odifreddi, P., Classical Recursion Theory, Vol. I, North-Holland (1989). 88. Peretyat’kin, M., Every recursively enumerable extension of a theory of linear orderings has a constructive model, Algebra i Logik, vol. 12 (1973), 211-219. 89. Pinus, A., Effective linear orderings, Siberian Math. J., vol. 16 (1975), 956-962. 90. Remmel, J. B., Recursively categorical linear orderings, Proc. Amer. Math. SOC.,V O ~ .83 (1981), 379-386. 91. Remmel, J. B., Recursive Boolean algebras, in Handbook of Boolean Algebras, Vol 3 (ed. D. Monk) North Holland (1990), 1097-1166. 92. Rice, H., Recursive real numbers, Proc. Amer. Math. SOC.,Vol. 5 (1954), 784- 79 1. 93. Rice, H., Recursive and recursively enumerable orders, Trans. Amer. Math. SOC.,V O ~ .83 (1956), 277-300. 94. Richter, L, Degrees of Structures, Ph. D. Thesis, University of Illinois at Urbana-Champaign, 1979. 95. Richter, L. J., Degrees of structures, J. Symb. Logic, vol. 46 (1981), 723731. 96. Rival, I., Ordered Sets (ed.) Nato Advanced Study Inst. Series 83, D. Reidel Co., Dordrecht (1981). 97. Rival, I., Graphs and Order (ed.) Nato Advanced Study Inst. Series, vol. 147, D. Reidel Co., Dordrecht (1984). 98. Rival, I., Combinatorics and Ordered Sets (ed.), Contemporary Math., vol. 57, Amer. Math. SOC.(1986). 99. Rogers, H. J., Theory of Recursive finctions and Effective Computability, McGraw-Hill, New York (1967). 100. Rosenstein, J., Linear Orderings, Academic Press, New York (1982). 101. Rosenstein, J., Recursive linear orderings, in Orders: Descriptions and Roles (M. Pouzet and D. Richard, eds.), Ann. Discrete Math., vol. 23, North-Holland (1982), 465-476. 102. Slaman, T., Relative to any nonrecursive set, Proc. Amer. Math. SOC,Vol 126 (1998), 2117-2122. 103. Slaman, T. and H. Woodin, Extending partial orders to dense linear orders, Ann. Pure and Appl. Logic, Vol. 94 (1998), 253-261. 104. Solomon, R., Reverse Mathematics and Ordered Groups, PhD. Thesis, Cornell University, 1998. 105. Solomon, R., Ordered groups: a case study in reverse mathematics, Bull. Symb. Logic, Vol. 5 (1999), 45-58. 106. Soare, R. I., Recursion theory and Dedekind cuts, Trans. Amer. Math. SOC.,V O ~ .140 (1969), 271-294. 107. Soare, R. I., Recursively Enumerable Sets and Degrees, Springer-Verlag, New York (1987). 108. Soskov, I., A jump inversion theorem for the enumeration jump, to appear,
102
Archive for Math. Logic. 109. Szpilrajin, S., Sur les tension de l’orde partial, Fund. Math., vol. 45 (1958), 2 13-2 16. 110. Schmerl, J., Recursion theoretical aspects of graphs and orders, in I. Rival [1984], 467-486. 111. Schmerl, J., What’s the difference?, Annals Pure and Applied Logic, Vol. 93 (1998), 255-261. 112. Smith, R., Two theorems on autostability in pgroups, in Logic Year 197980,Lerman, Schmerl, Soare, Springer, LNM, 859, (1981) 302-311. 113. Thurber, J., Degrees of Boolean Algebras, Ph. D. Thesis, University of Notre Dame, 1994. 114. Watnik, R., Recursive and constructive linear orderings, Thesis, Rutgers University (1980). 115. Wehner, S., Enumerations, countable structures and Turing degrees, Proc. Amer. Math. SOC.,Vol 126 (1998), 2131-2139.
103
TRIVIAL REALS
ROD G. DOWNEY* School of Mathematical and Computing Sciences Victoria University of Wellington New Zealand
DENIS R. HIRSCHFELDT~ Department of Mathematics University of Chicago
U.S.A. ANDRE NIES Department of Computer Science Auckland University New Zealand
FRANK STEPHAN~ Mathematical Institute Unaversity of Heidekberg Germany
n) 6 Solovay showed that there are noncomputable reals a such that H ( a H ( l n ) 0(1),where H is prefix-free Kolmogorov complexity. Such H-trivial reals are interesting due to the connection between algorithmic complexity and effective randomness. We give a new, easier construction of an H-trivial real. We also analyze various computability-theoretic properties of the H-trivial reals, showing for example that no H-trivial real can compute the halting problem. Therefore, our construction of an H-trivial computably enumerable set is an easy, injury-free construction of an incomplete computably enumerable set. Finally, we relate the H-trivials to other classes of “highly nonrandom” reals that have been previously studied.
+
*Supported by the Marsden fund of New Zealand. +Partially supported by NSF Grant DMS-02-00465. $Supported by the Heisenberg program of the Deutsche Forschungsgemeinschaft (DFG), grant no. Ste 96711-2.
104
1. Introduction Our concern is the relationship between the intrinsic computational complexity of a real and the intrinsic randomness of the real. Downey, Hirschfeldt, LaForte and Nies [8,9] looked at ways of understanding the intrinsic randomness of reals by measuring their relative initial segment complexity. (In this paper, “random” will always mean “1-random” ; see Section 2 for basic definitions.) Thus, for instance, if a and p are reals (in (0,l)),given as binary sequences, then we can compare the complexities of a and p by studying notions of reducibility based on relative initial segment complexity. For example, we define a < K p if K ( a 1 n ) K ( @1 n) 0(1), where we will be denoting classical Kolmogorov complexity by K . For prefix-free Kolmogorov complexity H , we define a < H p analogously. The goal of the papers [8,9] was to look at the structure of reducibilities like the above, and interrelationships among them, as a way of addressing questions such as: How random is a real? Given two reals, which is more random? If we partition reals into equivalence classes of reals of the “same degrees of randomness”, what does the resulting structure look like? The classic example of a random real is the halting probability of a universal prefix-free machine M , Chaitin’s R = CgEdom(M) 2-lu1. It is well-known that R has the property that a < H R for all reals a. A natural question to ask is the following: Given reals a < R ,6 (for R E { H , K } ) ,what can be said about the computational complexity of a and p measured relative to, say, Turing reducibility? For example, if we restrict our attention to computably enumerable (= recursively enumerable) reals, that is to the ones whose left cuts are computably enumerable, then being H-complete like R implies that the real is Turing complete. A natural guess would be that for all reals, if a < R p then a
<
<
+
<
+
+
105
proved that there are A! noncomputable reals p such that H ( P r n) < H(1”) +0(1).Solovay’s proof is in an unpublished manuscript, and is long and difficult. All known proofs of Solovay’s theorem use variations of his technique. In Section 3 we will give a new, short and easy proof of a strengthening of Solovay’s result that such noncomputable “H-trivial reals” exist. (Such a proof also appears in Vereshchagin [30].) To state an extension of this result, we need another triviality notion. Answering a question of KuCera and of van Lambalgen, KuEera and Terwijn [15] constructed a set X which is low for random. Here we say that X is low for random (also known as Martin-Lof-low) if the collection of sets random relative to X is exactly the collection of random sets. It is possible t o modify the construction given in Section 3 to show that there exist noncomputable computably enumerable sets that are both H-trivial and low for random. (Recently, Nies [23] has shown that in fact a real is H-trivial if and only if it is low for random.) H-triviality is surely a remarkable phenomenon. The remainder of the present paper is devoted to exploring this concept. We prove that no H-trivial real can be Turing complete, or even high. (Nies [23] has extended this result by showing that every H-trivial real is low.) An immediate application of this result is that the construction of a noncomputable H-trivial real provides a very simple injury-free solution to Post’s problem. Indeed, in Section 3 we give an alternate construction of a noncomputable H-trivial real that is not only injury-free but priority-free, in a sense that will be discussed in that section. We also prove that there is an effective listing of the H-trivial reals along with constants witnessing their H-triviality. This is in contrast to the strongly H-trivial reals, which have no such listing, as shown by Nies [24]. (A real is strongly H-trivial if H-complexity relativized to A is the same as H-complexity, up to an additive constant.) Recently, Nies and Hirschfeldt (see Nies [23]) have shown that a real is H-trivial if and only if it is strongly H-trivial. This was conjectured by Hirschfeldt and obtained as a direct modification of Nies’ result that the H-trivial reals are downward closed under Turing reducibility. (See the introduction to [23] for more details on the history of this result.) Our result shows that there is no computable way of passing from a constant witnessing H-triviality to one witnessing strong H-triviality. In an unpublished report, Zambella [31] proved that there is a computable function f such that for each c there are at most f (c) many reals
106
a with H ( a 1 n)
< H(ln) + c.
We will give a unified proof of this result and Chaitin’s result that every H-trivial real is A!. The reducibility < H is a preordering and hence we can form a degree structure from it, the H-degrees. The resulting degree structure on the computably enumerable reals has as its join operation ordinary addition. That is, [a] V [/?I = [a + /?I, where [a]is the H-degree of a. The H-trivial reals form the least H-degree. The study of relative randomness seems intimately related to weak truth table reducibility. Recall that A B if there is a Turing procedure @ and a computable function cp such that G B = A and for all x the maximum number queried of B on input x is bounded by cp(x). We prove that the H-trivial reals form an ideal in the wtt-degrees. Related to the topic of H-triviality is work of Kummer [lS]. Kummer investigated “Kummer trivial” computably enumerable sets. In terms of classical (non-prefix-free) Kolmogorov complexity, we know that if A is a computably enumerable set then K ( A 1 n) 2logn 0(1)for all n. Kummer constructed computably enumerable sets A and constants c such that, infinitely often, K ( A 1 n) 2 2logn - c. He called such sets complex. Kummer also showed that the computably enumerable degrees exhibit a gap phenomenon. Namely, either a degree a contains a complex set A, or all computably enumerable A E a are “Kummer trivial” in the sense that K ( A 1 n) 6 (1 6)logn O(1) for all E > 0. (By Chaitin’s work [3], if K ( A 1 n) logn 0(1)then A is computable, so this result is sharp.) Kummer proved that the degrees containing such complicated sets are exactly the array noncomputable (= array nonrecursive) degrees (see Section 7 for a definition). We prove that (I) no array noncomputable computably enumerable set is H-trivial, and (11) there exist Turing degrees containing only Kummer trivial sets which contain no H-trivial sets. The result (11) implies that being Kummer trivial does not make a set H-trivial.
<
<
+
+
+
+
2. Basic Definitions Our notation is standard, except that we follow the tradition of using H for prefix-free Kolmogorov complexity and K for non-prefix-free complexity. Following a recent proposal to change terminology, we call the recursively enumerable sets computably enumerable and the array nonrecursive sets
107
array noncomputable. The remaining computability-theoretic notation follows Soare's textbook [26]. We work with reals between 0 and 1, identifying a real with its binary expansion, and hence with the set of natural numbers whose characteristic function is the same as that expansion. A real a is computably enumerable if its left cut is computably enumerable as a set, or equivalently, if a is the limit of a computable increasing sequence of rationals. We work with machines with input and output alphabets {O,l}. A M ( r ' ) t for all machine M is prefix-free (or self-delimiting) if M ( T )4 finite binary strings T 7'. It is universal if for each prefix-free machine N there is a constant c such that, for all binary strings 7,if N ( T )4 then M ( / I ) $ = N ( T )for some /I with 1/1 IT) c. We call c the coding constant of N . For a prefix-free machine M and a binary string T , let H M ( T )be the length of the shortest binary string a such that M ( a ) J= , r , if such a string exists, and let H M ( T )be undefined otherwise. We fix a universal prefix-free machine U and let H ( T ) = H u ( T ) . The number H ( r ) is the prefix-free complexity of T . (The choice of U does not affect the prefix-free complexity, up to a constant additive factor.) For a natural number n, we write H ( n ) for H ( l n ) . A real a is random, or more precisely, 1-random, if H ( a 1 n) 3 n - O(1). There are several equivalent definitions of 1randomness, the best-known of which is due to Martin-Lof [21]. References on algorithmic complexity and effective randomness include Ambos-Spies and KuEera [l],Calude [2], Chaitin [4], Downey and Hirschfeldt [7], Fortnow [12], Kautz [14], Kurtz [17], Li and Vitanyi 1191, and van Lambalgen [29]. The above definitions can be relativized to any set A in the obvious way. The prefix-free complexity of a relative to A is denoted by H A ( a ) . An important tool in building prefix-free machines is the Kraft-Chaitin Theorem.
+
< +
Theorem 2.1. (Kraft-Chaitin) From a computably enumerable sequence of pairs ((ni,oi))iEw(known as axioms) such that 2-"' 1, we can effectively obtain a prefix-free machine M such that for each i there is a ~i of length ni with M(ri)$= ui, and M ( p ) t unless /I = ri for some i.
xiEw <
A sequence satisfying the hypothesis of the Kraft-Chaitin Theorem is called a Kraft-Chaitin sequence.
108
3. A short proof of Solovay's theorem We now give our simple proof of Solovay's theorem that H-trivial reals exist. This was proved by Solovay in his 1974 manuscript [27]. The proof there is complicated and only constructs a A: real.
Theorem 3.1. (after Solovay [27]) There is a noncomputable computably enumerable set A such that H ( A 1 n) 6 H ( n ) O(1).
+
Remark 3.1. While the proof below is easy, it is slightly hard to see why it works. So, by way of motivation, suppose that we were to asked to "prove" that the set B = (0" : n E w} has the same complexity as w = {In : n E w}. A complicated way to do this would be for us to build our own prefix-free machine M whose only job is to compute initial segments of B. The idea would be that if the universal machine U converges to 1" on input u then M ( a )$= 0". Notice that, in fact, using the Kraft-Chaitin Theorem it would be enough to build M implicitly, enumerating the length axiom (IuI,O"). We are guaranteed that x r E d o m ( M ) 2-IT1 6 CuEdom(U)2-1ul 6 1, and hence the Kraft-Chaitin Theorem applies. Note also that we could, for convenience and as we do in the main construction, use a string of length la1 1, in which case we would ensure that x T E d o m ( M ) 2-I.I < 1/2.
+
Proof of Theorem 3.1. The idea is the following. We will build a noncomputable computably enumerable set A in place of the B described in the remark and, as above, we will slavishly follow U on n in the sense that whenever U enumerates, at stage s, a shorter u with U ( a ) = n, then we will enumerate an axiom 1.1( 1 , A , 1 n) for our machine M . To make A noncomputable, we will also sometimes make A , 1 n # As+l 1 n. Then s, for the currently shortest string aj computing for each j with n 6 j j , we will also need to enumerate an axiom ( l a j l , A , + ~1 j) for M . This construction works by making this extra measure added to the domain of M small. We are ready to define A:
+
<
A = {(e, n) : 3s (We,,n A, = 0 A ( e , n ) E We,s A x 2 - H ( j ) [ s l < 2 - ( e + 2 ) ) } , (e,n)<j<s
where We,,is the stage s approximation to the eth computably enumerable set and H ( j ) [ s ]is the stage s approximation to the H-complexity of j. Clearly A is computably enumerable. Since &,,,2-H(j) goes to zero as m increases, if We is infinite then A[e]n Wpl
# 0. It is easy to see that
109
this implies that A is noncomputable. Finally, the extra measure put into the domain of M , beyond one half of that which enters the domain of U , is bounded by Ce2-(e+2) (corresponding to at most one initial segment change for each e ) , whence
Thus M is a prefix-free machine, and hence H ( A 1 n )
< H ( n ) + O(1).
a
We remark that the above proof can be modified to prove the result, which appears t o be due t o Muchnik, that there exists a noncomputable computably enumerable set A that is strongly H-trivial, in the sense that H complexity relativized to A is the same as H-complexity, up to an additive constant, Le., V(T(H((T) HA(o) c). Such an A is both H-trivial and low for random. (As mentioned above, Nies and Hirschfeldt (see Nies [23]) improved the result by showing that a real is H-trivial if and only if it is strongly H-trivial.) Clearly the proof also admits many variations. For instance, we can make A promptly simple, or below any nonzero computably enumerable degree. We cannot control the jump or make A Turing complete, since, as we will see, all H-trivials are nonhigh (and in fact, as shown by Nies [23], low). As we see in the next section, the construction above automatically yields a Turing incomplete computably enumerable set. It is thus an injuryfree solution t o Post's problem. It is not, however, priority-free, in that the construction depends on an ordering of the simplicity requirements, with stronger requirements allowed to use up more of the domain of the machine M . We can do methodologically better by giving a priority-free solution to Post's problem, in the sense that no explicit diagonalization (such as that of We above) occurs in the construction of the incomplete computably enumerable set, and therefore the construction of this set (as opposed to the verification that it is H-trivial) does not depend on an ordering of requirements. We now sketch this method, which is rather more like that of Solovay's original proof of the existence of a A: H-trivial real. Let us reconsider the key idea in the proof of Theorem 3.1. At certain stages we wish t o change an initial segment of A for the sake of diagonalization. Our method is to make sure that the total measure added to the domain of our machine M (which proves the H-triviality of A) due to such changes is bounded by 1. Suppose, on the other hand, we were fortunate in
<
+
110
the sense that the universal machine itself “covered” the measure needed for these changes. That is, suppose we were lucky enough to be at a stage s where we desire to put n into A,+1 - A , and at that very stage H , ( j ) changes for all j E {n,. . . ,s } . That would mean that in a n y case we would need to enumerate new axioms describing A,+I f j for all j E { n ,. . . ,s } , whether or not these initial segments change. Thus at that very stage, we could also change A , f j for all j E {n,. . . ,s} at no extra cost. Notice that we would not need to copy the universal machine U at every stage. We could also enumerate a collection of stages t o , tl . . . and only update M at stages ti. Thus for the lucky situation outlined above, we would only need the approximation to H ( j ) to change for all j E {n,. . . ,t,} at some stage u with t, u ts+l. This observation would seem t o allow a greater possibility for the lucky situation to occur, since many more stages can occur between t, and ts+l. The key point in all of this is the following. Let t o , t l , . . . be a computable collection of stages. Suppose that we construct a set A = Ata so that for n 6 t,, i f Ats+, f n # At8 f n then His ( j ) < Hts+l( j ) for all j with n j t,. Then A is H-trivial. We are now ready to define A in a priority-free way. Let to,tl,. . . be a collection of stages such that ti as a function of i dominates all primitive recursive functions. (Actually, as we will see, dominating the overhead in the Recursion Theorem is enough.) At each stage u,let {ai,, : i E w } list &. Define
< <
u,
< <
Ats+, = At* u
. . ,t s } ,
<
where n is the least number t, such that H t S + , ( j ) < H t , ( j ) for all j E { n , .. . t s } . (Naturally, if no such n exists, Ats+, = A t e . ) Requiring the complexity change for all j € { n , .. . ,t s } , rather than just j E {an,ts,.. . ,t,}, ensures that A is coinfinite, since for each n there are only finitely many s such that (n) < Hts (n). Note that there is no priority used in the definition of A . It is like the Dekker deficiency set or the so-called “dump set” (see [26], Theorem V.2.5). It remains to prove that A is noncomputable. By the Recursion Theorem, we can build a prefix-free Turing machine M and know the coding constant c of M in U . That is, if we declare M ( u ) = j then we will have U ( T ) = j for some T such that 171 6 IuI c. Note further that if we put u into the domain of M at stage t,, then r will be in the domain of U by stage ts+l - 1. (This is why we chose the stages to dominate the primitive recursive functions. This was the key insight in Solovay’s original
+
111
construction.) Now the proof looks like that of Theorem 3.1. We will devote 2-e of the domain of our machine M to making sure that A satisfies the e-th simplicity 2-Ht*(j) < requirement. When we see an,ts occur in We,t,, where CnCjCt, 2-(e+c+1), we change the Mta descriptions of all j with n 6 j 6 t, so that Ht.+l (j) < Hi8( j ) for all such j . The cost of this change is bounded by 2 - e , and a,,ts will enter Ata+,, as required. 4. Turing degrees of H-trivials
In this section we give a proof that every H-trivial real a is Turing incomplete, and in fact not high (i.e., a’
Theorem 4.1. If the real A is H-trivial then A is not high. Proof. By Corollary 6.2 part (b), we can choose a A: approximation for A. We enumerate a Kraft-Chaitin sequence L of axioms of the form ( r n ,n). If E N, the weight of E is CnEN 2-rn. Our enumeration of L depends on the behavior of a total Turing reduction rA.At first we will be quite general about what r is, but we adopt the usual conventions on the use y of r. The basic idea is as follows. We enumerate an axiom (r,n) into L for some small r , thus ensuring that H(n) is small. Each time A 1 n changes, the “opponent” has to provide a corresponding short description of A 1 n via the fixed universal prefix-free machine U . Thus we make the opponent load up many descriptions of approximations to A 1n, while we enumerate only one short description of n. We will be able to argue that there are enough A n-changes for certain n, which will be picked so that n > ~ ( r n ) for certain numbers rn. We first consider the simpler case where we only want to show that A is Turing incomplete. Fix a Turing reduction K = !PA. We build a computably enumerable set B. By the Recursion Theorem, we can assume that we know in advance a Computably enumerable index i for B. Since B equals the i-th row of K , this means we also know a total reduction r such that rA= B . (For the full construction, we will use the fact that if 0”
r
<
+
112
+
<
+
ahead of time. Let c = b d, so that V n ( H ( A n ) H . M ( ~ )c). All values appended by [s]will be taken at the end of the stage. We carry out our construction at stages so < s1 < . . . so that V i V n si ( H ( A f n)[si+l] ~ M ( n ) [ s i + l ] c ) and V i V n si ( r A ( ~ ) [ s i + 1 ] - 1 ) . Let k E N be a sufficiently large number (the precise value will be determined below). We describe procedures Pi(g), 0 i < k, where the parameter g is a rational of the form 2-”, 2 E N. Each procedure Pi, 0 i < k - 1, calls Pi+l many times. The basic idea is that, for each n we work with, each level of procedures is responsible for creating a change in A n, thus forcing the opponent to provide a short description of a new string of length n. For certain n, however, changes will come too early, thus creating a certain amount of “trash”, that is, axioms enumerated into L that do not cause the appropriate number of short descriptions to appear in U . We will need to show that this trash is small enough that it will not cause us problems.
<
+
<
<
<
<
The bottom procedure Pk-l(g). The output of this procedure (if the procedure returns at all) is a set c = c k - 1 of numbers n such that (I) (11)
first we put an axiom (rnln ) into L (and therefore have M ( c ) = n for some c of length r n ) ]and then we see strings eo,c1 of length n with U-descriptions of length rn c. (The strings are approximations to A n at certain stages.)
< +
of C equals g, and the “trash” put into Moreover, the weight CnEC2-Tn L by this procedure, namely CnBC 2-Tn, is at most 2-2(k-1)g. Here is the procedure:
(1) Choose a fresh number m = m k - 1 . (2) Let q k - 1 = 0 and C = 0. WHILE q k - 1 < 9: (a) Choose a fresh number n (in particular, n > ~ ( m )and ) put an axiom ( r , n ) into L, where T is given by 2-T = 2-2(k-1)2-(”+1)g, and u is the number of times the expression A t ~ ( mhas ) changed so far. (b) At the next stage (in the sense described above): IF A f ~ ( mchanged, ) add 2-T to the real trash (which is global for the construction, being 0 initially), and continue at 2(a>*
113
ELSE put n into C and add 2-' to qk-1. END WHILE (3) Wait for an A 1 ~ ( mchange. ) If this happens, RETURN the set C. (In the case where we only want to show that A is not Turing complete, we are enumerating B , and we know B = FA. Thus we may simply put m into B . For the proof of the full theorem, it remains to show that this A-change happens often enough.) This completes the description of the procedure. Note that the procedure can only get stuck at step 3, because we assume A 1 ~ ( msettles. ) We verify the required property (11) of C. If n gets into C at a stage s, then ~ ( m<) n (and we see a short description of A[s] n). Since we assume the procedure returns at a stage t > s, A 1 ~ ( mhas ) changed from its value at s, so at t there is a short description of a second string of length n. Clearly CnEC 2-Tn = g , since g has the form 2-" and we stop when qk-1 reaches this value. Moreover, trash ,< 2 - 2 ( k - 1 ) gC,,,, 2-("+1) \ < 2-2("-l)g.
r
<
The procedure P i ( g ) for 0 i < k - 1. The output of this procedure (if the procedure returns at all) is a set C = Ci of numbers n such that
(I) first we put an axiom (r,,n) into L , and (11) then at later stages we see strings oj (0 ,< j k - i 2 ) of length n with U-descriptions of length T , + c. (Again, these strings are approximations to A 1 n at those stages.)
<
<
+
Moreover, the weight C n E C 2 - T n of C is g , and the trash put into L by this procedure, namely CneCi 2-Tn, is at most 2-2ig. Here is the procedure: (1) Choose a fresh number m = mi. ( 2 ) Let qi = 0 and C = Ci = 8. WHILE qi < 9: (a) Call p i + l ( g i + l ) , where gi+l = 2-2(i+1)2-("+l)g, and 21 is the ) changed so far. number of times the expression A 1 ~ ( mhas If A 1 r(m)changes during the execution of Pi+l, put the current Cj,i q j into trash and end the subprocedures called by Pi.
114
(b) If Pi+l returns a set Ci+l then put this set into Ci, and add Si+l to qi. END WHILE (3) Wait for an A 1 ~ ( mchange. ) If this happens RETURN the set C. This completes the description of the procedure. Again, we verify the required property (11) of Ci. When Pi+l returns a set Ci+l at stage s , by induction, for each n E Ci+l we have already seen short descriptions of distinct strings aj (0 6 j < k - i + l ) of length n. Since this run of Pi+l was not stopped, A 1 $mi) did not change during this run, and in particular $mi) < n. If A f mi) changes before Pi returns, this gives a new string of length n with a short description. Otherwise A mi) changes when Pi leaves step 3, again giving a new string of length n with a short description. Lemma 4.1. If P i ( g ) is called and stopped during the loop performed at step 2, then the amount added t o trash by P i ( g ) and the subprocedures it calls is a t m o s t 2pZig.
Proof. We use induction on descending values of i. Above, we verified the lemma for i = k - 1. Now suppose the statement is true for i 1. (a) If P i ( g ) calls Pi+l at stage s with goal gi+l,s, then by the induction hypothesis, a t most 2-2(i+1)gi+l,s is put into t r a s h by this subprocedure. Since C s g i + l , s gi, the total contribution to trash of the subprocedures called by c ( g ) is 2-2(i+1)gi. (b) When Pi stops its subprocedures during execution of step 2 , it puts C j > i qj into t r a s h . But always qj+l q j / 2 , and letting u be the number of times the expression A f ~ ( mhas ) changed so far, qi+l < , 2-2i-22-(~+1) 9. So Cj >.a . q3' < 2-2i-12-(U+1)g, and the total sum over all u is 2-2i-1g. The trash contributed by (a) and (b) together is at most 2-2ig. 0
+
< <
<
<
The proof of Turing incompleteness of A runs as follows. Let k = 2c+3. We start the construction by calling Po(g) with g = 114. Then CnECo 2-'- = g and, by Lemma 4.1, trash g . Thus the total weight put into L (i.e., the weight of N) is 6 2g, and hence L is a Kraft-Chaitin sequence. Now, by induction on descending i < k , we can show that each run of a procedure P i ( g ) returns unless stopped. Suppose i = k - 1, or i < k - 1 and the claim is true for i 1. Since FA is total, eventually the counter u in step 2 is constant. So if i < k - 1 then eventually we call Pi+I(g/) for a fixed g' often enough to reach the required weight (for i = k - 1 the
<
+
115
argument is similar). We can enforce the A 1 ?(mi) change needed at step 3 by enumerating mi into B . Thus Pi(,) returns. When the initial procedure Po(g)returns, by the property (11) of Co,the opponent has to provide at least measure kg2-c in descriptions of length n strings. So if g = 1/4, we reach the contradiction p(dom(U)) > 1. We now complete the Proof of Theorem 4.1. Assume 0 '
j > i. We want to prove that only finitely many runs R, get stuck at step 3. Let the total computable function f i ( x ) be defined as follows. If x is not the parameter mi of a run R, with v 2 vo by stage x, then f(x) = 0. Otherwise, fi(x) is one greater than the first stage where mi has been canceled or y(z) is defined and R, reaches step 3. (Such a stage exists by the induction hypothesis and the fact that I? is total.) Now for almost all z, the initial segment A y(z) changes after the stage when f(x) has been defined, so the corresponding run R, does not get stuck. 0
116
Thus, for a sufficiently large p , we complete procedures Po(2--P) as many times as needed t o reach a set COof weight 1/4. This gives a contradiction as before. One final limitation is the following.
Theorem 4.2. The Turing degrees of H-trivial reals are bounded by a computably enumerable degree strictly below 0’. Proof. Nies (unpublished) has shown that every H-trivial real is Turing reducible t o an H-trivial computably enumerable set, so it is enough to prove the theorem for the degrees of H-trivial computably enumerable sets. Notice that the statement “H(Wi 1 n) H ( n ) c for all n” is II; in the parameters i, c. Thus the collection of indices of H-trivial computably enumerable sets is E!. We can enumerate a piecewise computably enumerable set A where the (i,c)-th column is equal to Wi iff Wi is H-trivial with constant c, and finite otherwise. By Theorem 7.1, the H-trivials are closed under join, so such a set has the property that $m6nA(m) is Turing incomplete for all rn. Hence, the result follows from the strong form of the Thickness Lemma (see Soare [26], Ch. VIII, Theorems 2.3 and 2.6).
<
+
In unpublished work, Nies has shown that the degrees of H-trivial computably enumerable sets are bounded below 0’ by a low2 computably enumerable set. This is much more difficult to prove. 5. Listing the H-trivials We next prove a result about the presentation of the class 31 of H-trivial reals. First consider the computably enumerable case. As is true for every class that contains the finite sets and has a .E! index set, there is a uniformly computably enumerable listing ( A , ) of the computably enumerable sets in 3t. Here we show there is a listing that includes the witnesses of the Cg statement, namely the constants via which the A, are H-trivial. This is true even in the A; case. We say that a is H-trivial via the constant c if H ( a n) H(n)+c for all n. A A:-approximation is a computable ( 0 , 1)-valued function Xz, s B,(z) such that B,(z) = 0 for z 2 s and B ( z )= lim, B,(z) exists for each z.
<
Theorem 5.1. There is a n eflective list ( ( B e , s ( z ) ) S E N , d e ) of A:approximations and constants such that each H-trivial real occurs as a real Be = lim, Be,,, and each Be is H-trivial via the constant d,. Moreover,
117
B,,,(x) changes at most O ( x 2 )times as s increases, with effectively given constant. Proof. We define, uniformly in e , A:-approximations Be,S and KraftChaitin sequences Ve such that, for effectively given constants g, and for each stage u,
< H s ( n ) + ge + 3 ( ( r ,B e , s t n) E Ve,s). (1) Then we obtain d, by adding to g, + 2 the coding constant of a prefix-free Vw
<
1 ‘1
3r
machine uniformly obtained from V,. We need a lemma whose proof will be obtained by analyzing the proof of Theorem 5.8 in [23]. For those familiar with that paper, we include a proof of this lemma below. The lemma says that there is a uniformly computable set Q , of “good stages” such that Be changes only at a good stage, and the cost of these changes, namely the weight of short descriptions of the new initial segments Be,s1 rn, is bounded by an effectively given constant 29..
Lemma 5.1. There is a n effective list ((B,,s(x)),EN,ge,Qe) of A:approximations, constants, and (indices for) computable sets of stages, with the following properties.
*
(1) Be,u(x) # Be,u-l(X) 21 E Q e . (2) Let Qe = ( q e ( 0 ) < qe(1) < * * * } ( Q e m a y be finite). If qe(r defined, then let t ( z , r ) = Cz
S, =
{ t ( z , r ): u = qe(rf 2) defined
A
x is minimal such that B,,U(x) Then S e
+ 1) is
# Be,U-l(x)}.
< 2ge.
Moreover, B,,,(x) changes at most O ( x 2 ) times as s increases, with effectively given constant. We first complete the proof of the theorem assuming the lemma. We obtain V, by emulating the construction of an H-trivial real in Theorem 3.1 (see also [23], Proposition 3.3). At stage u,for each w u,put (HU(w)g, 3, B, w) into V, in case
<
(a) u = w,or (b) u > w A HU(w) < HU-1(w), or (c) B U - 1 t w # B U t w.
+ +
118
Clearly, each V, satisfies (1). It remains to show that V, is a Kraft-Chaitin sequence. We drop the subscript e in what follows. The weight contributed by axioms added for reasons (a) and (b) is at most 2-9-2 6 1/4. Now consider the axioms added for reason (c). Since B only changes at stages in Q, for each w there are at most two enumerations at a stage u = q(r 2 ) such that w > q ( r ) . The weight contributed by all w at such stages is at most R/4. Now assume w q ( r ) ,and let u = q(r 2 ) . Case 1. Hq(,+l)(w) > H,(w).This happens at most once for each value H,(w), u E Q. Since each value corresponds to a new description of w, the overall contribution is at most n/8. Case 2. Hq(T+l)(w) = H,(w). Since B ( z ) changes for some minimal z < w at u , the term 2-Hu(w) occurs in the sum e ( z , r ) .Since S 29, the 0 overall contribution is at most 1/8.
+
<
+
<
Proof of Lemma 5.1. By [23], Theorem 6.2, let (l?m)mEN be a list of (total) tt-reductions such that the class of H-trivial reals equals {Fm(O’): m E N}. Let A, = I’,(0’), with the A!-approximation A,+ = rm(O:). We refer to the proof of [23],Theorem 5.8, and adopt its notation. Let e be a computable code for a tuple consisting of the following: rn, a constant b (we hope A, is H-trivial via b), numbers i (a level in the tree of runs of procedures) and n (we consider the n-th run of a procedure Pi(p,a),hoping it will be golden), and a constant g, which we hope will be such that 29. = p / a . (We assume that ge is at least the constant via which the empty set is H-trivial.) Given e, we define a set Qe. If e meets our expectations then Qe will be equal to A, and will be H-trivial via ge. Otherwise, Q , will be finite, but g, will still be a correct constant via which Q , is H-trivial. As in the main construction, we obtain a coding constant d for a prefixfree machine by applying the Recursion Theorem with parameters to rn, b, let k = 2b+d, and only consider those i 6 k. Given e, we run the construction as in [23], Theorem 5.8, in order to define Q,. For each u,we can effectively determine if u is a stage in the sense of that construction. Moreover we can determine if by stage u we started the n-th run Pi(p, a) of a procedure Pi. We leave Qe empty unless ge = p / a . In that case we check if u = q ( r ) in the sense of [23], Theorem 5.8. If so we declare u E Qe. Finally we let Be,,(z) = A,,,,,(~~n(o,...,~}).Thus if Q , is finite we are stuck with Am,m,xQe. The property S 6 29. is verified in the proof of [23],Theorem 5.8. The O ( x 2 )bound on the number of changes follows as
119
in [23], Fact 3.6.
0
Note that we can replace the list ),'I( in the above proof by a listing of a subclass of the H-trivials containing the finite sets. Thus there are also effective listings with constants for the H-trivial computably enumerable sets and for the H-trivial computably enumerable reals. Let C be a set of computably enumerable indices closed under equality of computably enumerable sets. We say that C is uniformly E: if there is a ll! relation P such that e E C H 3n ( P ( e ,n ) ) and there is an effective sequence (en,bn) such that P(e,,b,) and V e E C3n(We = W e , ) . We have proved that 3t is uniformly E:. It would be interesting to see which other properly E : index sets have that property, for instance the class of computable sets. Recall that A is strongly H-trivial via a constant c if V a ( H ( o ) H A ( a ) c ) , where HA is H-complexity relativized to A. In [23] it is proved that each H-trivial real is strongly H-trivial. However, in the proof the constant of strong H-triviality is not obtained in a uniform way. The following corollary shows that this non-uniformity is necessary.
+
<
Corollary 5.1. There is no effective way t o obtain from a pair ( A , b ) , where A is a computably enumerable set that is H-trivial via b, a constant c such that A is strongly H-trivial via c. Proof. Otherwise, by Theorem 5.1 above we would obtain a listing (Ae,c e ) of all the strongly H-trivials with appropriate constants. Nies [24], Theorem 0 5.9, showed that such a listing does not exist. 6. Theorems of Chaitin and of Zambella
In this section we give a unified proof of some unpublished material of Zambella and of Chaitin's result that all H-trivials are A;, while establishing some intermediate results of independent interest.
Definition 6.1. Given a prefix-free machine D ,let Z o ( a ) = p ( D - ' ( u ) ) . That is, Zo(a)is the probability that D outputs g. If D is the fixed universal machine we will write Z ( G )for Z o ( a ) . Theorem 6.1. ZD(G)= 0 ( 2 - H ( " ) ) . We make a few comments before proving this theorem. A measure of complexity is any function F : 2<w + w such that x,2-F(") < 1
120
<
and {(o,Ic) : F ( o ) k} is computably enumerable. Chaitin [3] introduced this concept and showed that H-complexity is a minimal measure of complexity in the sense that, for any measure of complexity F , we have H ( o ) F ( o ) O(1). Notice that - log, Z(a) is a measure of complexity, and hence, by the minimality of H among measures of complexity, we know Z(o). Therefore, by Theorem 6.1, we know that for some that 2-H(") constant d ,
<
+
<
2TH(")
< Z(a) < d2TH(").
Thus we can often replace usage of H by 2. As an illustration, for reals and p, we have the following result.
Theorem 6.2.
Q \
Q
p iff there is a constant c such that for all n, CZ(P t
2 Z(Q t n ) .
Proof. Suppose that Q
+
2-H(aIn) 2 d12-H(0 In).
This happens iff there is a c such that Z ( a 1 n) 2 cZ(P 1 n) for all n. The other direction is similar. 0
Remark 6.1. For any o,the real Z(a) is random. To see that the remark is true we use the Kraft-Chaitin Theorem to build a machine M and show that R 6 s Z(a),where <S is Solovay reducibility (see [9] for a definition and discussion of Solovay reducibility). At stage s, if we see U ( v )4,where U is the universal machine, we declare that M ( v ) = o. Then for some c = C M , there is a v' with U ( v ' ) = o, and furthermore IvI lv'l c. Thus whenever we add 2-14 to R, we add 2-(l"l+c) to Z ( a ) , and hence 52 6 s Z(o), which implies Z(o) is random.
<
+
Proof of Theorem 6.1. The idea of the proof is the following: We will use the Kraft-Chaitin Theorem to define a prefix-free machine M as follows. Whenever we see Zo(a) 2 22r-n, where n is the current H-complexity of (T,we will enumerate an axiom (n - T 1,o) (saying that some string of length n - T 1 is mapped to o by M ) . For large enough T we will get to contradict the minimality of H. In detail, at stage s, we do the following. For each u , n , < ~ s, if
+
+
121
0
u,n , s is not yet attended to, n>2rb2, H ( o ) [ s ]= n , and Zo(u) 2 22'--n,
+
then attend to o , r , n by enumerating an axiom ( n - r 1,u). Notice that for any fixed u,r , we put in axioms ( n - r 1,u) for descending values of n . Let ho,r be the last value put in. We add at most
C 2-(hu,~-T+l+n)=
+
00
2-h,7-+r
n=O
to the measure of the domain of M . When we put in the last axiom (h,,,' - r 1,o),we see that Zo(a) 2 22T-hu9r. Since D is prefix-free, for this fixed r we can conclude that
+
c
22r--h,-,
< 1.
U
Therefore,
Hence, for r we can add at most 2-' to the measure of the domain of M . Thus, as r 2 1, we can apply the Kraft-Chaitin Theorem to conclude that M exists. Let c be such that
Let d = 22(c+2).Then we claim that
To see this, let r = c+2. If Zo(u) > 22'-H(u), then eventually we put in an axiom ( H ( u ) - r + l , u ) , a n d h e n c e H ~ ( o ) H(o)-(c+l),acontradictioa
<
This result allows us to get an analog of the result of Chaitin [3] on the number of descriptions of a string. Corollary 6.1. There is a constant d such that for all c and all u,
I{.
:D(v)= u A
1.1 6 H ( u ) + c}l
< d2".
122
Proof. Trivially, p ( { v : D ( v ) = a A IvI
< H ( n )+ c } ) 2 2-(H(")+c) *
{I.
But also, p ( { v : D ( v ) = a A (vI f H ( a ) 6.2. Thus,
:D(v) =
A
1 ~ 61 H ( a ) +.}I.
+ c } ) 6 d . 2 - H ( u ) , by Theorem
d2-H(") 2 2 - C 2 - H ( u ) l { ~: D ( v ) = a A lvl f H ( a ) +.}I. Hence, d2c 2 I{v : D ( v ) = a A IvI f H ( a ) +.}I.
0
We can now conclude that there are few H-trivials.
Theorem 6.3. The set s d = { a : H ( a ) < H(Ia1) many strings of length n.
+ d } has at most O ( 2 d )
Proof. Given a universal prefix-free machine U , there is another machine V with the following property: V has for each n a program of length m (on which it converges) whenever the sum of all 2-IPI such that U ( p ) is defined and has length n is at least 2lPrn;furthermore V has for every n and every length m at most one program of length m. As U is universal, it follows that there is a constant c such that the following holds: If the sum of all 2-IP such that U ( p ) is defined and has length n is at least 2C-m, then there is a program q of length m with U ( q ) = n. Let m = H ( n ) and n be any length. There are less than 2d+c+1 many programs p of length m + d or less such that U ( p )has length n, as otherwise the sum 2-IPI over these programs would be at least 2C+1-rn, which would cause the existence of a program of length m - 1 for n, a contradiction to H ( n ) = m. So the set s d = {a : H ( a ) < H(Ia1) d } has at most 2d+c+1 many strings of length n, where c is independent of n and d. 0
+
Corollary 6.2. (a) (Zambella [31]) For a fixed d , there are at most 0 ( 2 d ) many reals a with
H ( a r n) 6 H ( n )
+d
for all n.
(b) (Chaitin [3]) If a real is H-trivial, then it is A:. Proof. Consider the A: tree T d = {a : Vv a ( v E s d ) } . This tree has width O ( z d ) ,and hence it has at most 0 ( 2 d )many infinite paths. For each such path X , we can choose a E T d such that X is the only path above a. Hence such X is A:. 0
123
7. Triviality and wtt-reducibility Recall that A G w t t B iff there is a procedure @ with computable use cp such that G B = A. As we have seen in the earlier papers mentioned in the introduction, wtt-reducibility seems to have a lot to do with randomness considerations. Triviality is no exception.
Theorem 7.1. Suppose that a trivial.
<,,,ti
,8 and ,B is H-trivial. Then a is H -
Proof. For each computable cp : N H
N,
H(cp(n)) 6 H ( n ) + W ) .
(To see this consider the prefix-free machine M such that for all o,if U ( o ) = n then M ( o ) = cp(U(o)),where U is a universal prefix-free machine.) Now suppose that a = with computable use cp and that p is H trivial. We have H ( a t n) 6 H(P t cp(n))+ 00) 6 H(cp(n)) + O(1) 6 H ( n ) + W ) , by the above.
0
Nies [23] has extended this result to Turing reducibility, but with a much more difficult proof. We now show that the H-trivials are closed under join, and hence form an ideal in the wtt-degrees. We begin by showing that the H-trivials are closed under addition.
Theorem 7.2. If a and ,f3 are H-trivial then so is a + p. Proof. Assume that a , P are two H-trivial reals. Then there is a constant c such that H ( a 1 n) and H ( P 1 n) are both below H ( n ) c for every n. By Theorem 6.3 there is a constant d such that for each n there are at most d strings T E (0, l}n satisfying H ( T ) H ( n ) c. Let e be the shortest program for n. One can assign to a t n and /3 n numbers i , j 6 d such that they are the i-th and the j-th string of length n enumerated by a program of length up to [el c. Let U be a universal prefix-free machine. We build a prefix-free machine V witnessing the H-triviality of a P. Representing i,j by strings of the fixed length d and taking b E (0, l}, V ( e i j b )is defined by first simulating U ( e ) until an output n is produced and then continuing the simulation in order to find the i-th and j-th string a and p of length n such that both
+
<
+
+
+
124
+
are generated by a program of size up to n c. Then one can compute 2-n(a p + b) and derive from this string the first n binary digits of the real a p. These digits are correct provided that e, i , j are correct and b is the carry bit from bit n + 1 to bit n when adding a and fi - this bit is well-defined unless a + ,B = z . 2-m for some integers m, z , but in that case a ,8 is computable and one can get the first n bits of a ,8 directly without having to do the more involved construction given here. 0
+ +
+
+
Corollary 7.1. The wtt-degrees containing H-trivials form an ideal in the wtt-degrees.
Proof. By Theorem 7.2, we know that if a and p are H-trivial, then SO is a p, where is normal addition. Now let a' = a(O)Oa(l)O . . . , where a(.) is the nth bit of a , and let p' = Op(O)Op(l). . .. Both a' and fi' are H-trivial, since they have the same wtt-degrees as a and p, respectively. It follows that a' p' = a @ ,O is H-trivial. 0
+
+
+
Theorem 7.2 suggests the question of whether addition is a join on the H-degrees. In general, it is not, as the example R g~ R (1 - R) shows. But for computably enumerable reals it is. This fact considerably simplifies the analysis of the H-degrees of computably enumerable reals (compare for instance the difficulties in studying the sw-degrees considered in IS], many of which arise from the lack of a join operation).
+
Theorem 7.3. If a , P are computably enumerable reals then the H complexity of Q /3 is - up to an additive constant - the maximum of the H-complexities of a and p. In particular, a /3 represents the join of a and O, with respect to H-reducibility.
+
+
Proof. Let y = a + p. Without loss of generality, the reals represented by a , p are in (0,1/2), so that we do not to have to care about the problem of representing digits before the decimal point. Furthermore, we have programs i, j , k which approximate a , p, y, respectively, from below, such that at every stage and also for the limit the equation a + /3 = y holds. First we show that H ( y n) max{H(a n),H(P n ) } c for some constant c. Fix a universal prefix-free machine U . It is sufficient to produce a prefix-free machine V that for each n computes ( a p) 1 n from some input of length up to max{H(a n),H(,8 1 n ) )+ 2. The machine V receives as input eab where a, b E {0,1} and e E (0, l}*. The length of the input is lel + 2. First V simulates U ( e ) . In the case
<
r
r
+
+
125
that this simulation terminates with some output (T,let n = TI. Now V simulates the approximation of a and p from below until it happens that either 0
0
a=Oanda=arnor a=landa=p/n.
Let 6 ,b be the current values of the approximations of a and p, respectively, when the above simulation is stopped. Now V outputs the first n bits of the real ii b . 2-n. In order t o verify that this works, given n, let a be 0 if the approximation of p is correct on its first n bits before the one of a and let a be 1otherwise. Let e be the shortest program for a 1 n in case a = 0 and for ,f3 1 n in case max{H(a 1 n ) , H ( P 1 n ) } . In a = 1. Then U ( e ) terminates and lel addition, we know both values a 1 n and ,f3 1 n once U ( e ) terminates. So 6 and ,6 (defined as above) are correct on their first n bits, but it might be that bits beyond the first n cause a carry to exist which is not yet known. But we can choose b t o be that carry bit and have then that V ( e a b ) = y 1 n. For the other direction, we construct a machine W that computes (a 1 n,P 1 n ) from any input e with U ( e ) = y 1 n. The way to do this is t o simulate U ( e ) and, whenever it gives an output (T,to simulate the enumerations of a, p, y until the current approximation 1 n = (T. As 6 6 = ;U, it is impossible that the approximations of a , ,d will later change on their first n bits if y / n = o. So the machine W then just outputs (6 1 n,,8 1 n ) ,which is correct under the assumption that e, and therefore also (T,are correct. 0
+ +
<
+
Recall that a computably enumerable set X is (Kummer) complex iff K ( X / n) 3 2logn - c infinitely often. (No computably enumerable set can have K ( X 1 n) 3 2logn - c for all n; see [19], Exercise 2.58.) Recall also that, by [16], a computably enumerable degree d either has complex computably enumerable sets or every computably enumerable set D E d is Kummer trivial in the sense that for all e > 0 there is a constant c such that for all n,
K ( D 1 n)
< (1+ E ) logn + d.
The relevant degrees containing the complex sets are the array noncomputable degrees of Downey, Jockusch and Stob. Recall that a very strong array {F, : z E N} is a strong array such that IF,I < IF,+11 for all z. A computably enumerable set A is called array noncomputable relative to
126
such a very strong array if for all computably enumerable sets W there are infinitely many x such that W n F, = A n F,. A relevant fact for our purposes is the following. Theorem 7.4. (Downey, Jockusch and Stob [lO,ll])For all w t t degrees d, and all very strong arrays {F, : x E N}, if d contains a set that is array noncomputable relative t o some very strong array, t h e n d contains one that is array noncomputable relative t o {F, : x E N}.
We first show that array noncomputable wtt-degrees (i.e., ones containing array noncomputable computably enumerable sets) cannot be H-trivial. Theorem 7.5. If d is a n array noncomputable and computably enumerable
wtt-degree t h e n n o set in d is H-trivial. Proof. We will build a prefix-free machine M . The range of M will consist of initial segments of lW.By the Recursion Theorem, we can assume we know the coding constant d of our machine in the universal prefix-free machine U . Choose a very strong array such that IF,I = 2d+e+1. By Theorem 7.4, d contains a set A array noncomputable relative to this array. We claim that A is not H-trivial, and hence the wtt-degree d contains no H-trivials. Suppose that A is H-trivial and that A < H lW with constant c. We will build a computably enumerable set V with V n Fg # A n Fg for all g > c, contradicting the array noncomputability of A. For each el we do the following. First we “load” 2d+e+1 beyond max{a: : a: E F,+,}, by enumerating into our machine an axiom (2d+ef1,1z) for some fresh t > max{z : 2 E F,+,}. The universal machine must respond at some stage s d + e + 1 + c. We by converging to A, 1 z on some input of length then enumerate into V,, our “kill” computably enumerable set, the least p E F,+, not yet in V,, making F,+, n A[s]# V, n F,+,[s].Notice that we only trigger enumeration into V at stages after a quantum of 2e+1+c+d has been added t o the measure of the domain of U . Now the possible number of changes we can put into V for the sake of e c is IF,+,I, which is bigger than 2e+c+1+d. Hence A cannot respond each time, since if it did then the domain of U would have measure bigger than 1. 0
<
+
One might be tempted to think that the Kummer trivial computably enumerable sets and the H-trivial sets correspond. The next result shows that at least one inclusion fails. Since the proof is rather technical, and the
127
result also follows from the recent results of Nies [23] mentioned above, we restrict ourselves to a brief proof sketch. Theorem 7.6. There is a computably enumerable Turing degree a that consists only of Kummer trivials but contains no H-trivials.
Proof sketch. We construct a contiguous computably enumerable degree a containing no H-trivials. A contiguous degree is a Turing degree that consists of a single wtt-degree. Such degrees were first constructed by Ladner and Sasso [18]. By Downey [ 5 ] , every contiguous computably enumerable degree is array computable. Hence, by Kummer's theorem, all of its members are Kummer trivial. The argument is a II! priority argument using a tree of strategies. We must meet the requirements below:
'Re.+ : @f= B
A
@f= A implies A E
, ~B ~
There is a standard way to do this via dumping and confirming. Specifically, one has a priority tree PT = ( 0 0 , f } ' w with versions of R,,i having outcomes 00 < L f. The outcome 00 is meant to say that @! = B A @? = A. The other outcome is meant to say that the limit of the length of the agreement function e(e,i,s) = max{z : Vy f z(@f(y) = A(y) A Vz
< c p ( y ) ( @ t ( z=) B ( z ) [ s ] ) ) } ,
the so-called A-controllable length of agreement, is finite. The usual H-nontriviality requirements are that
P, : A is not H-trivial via e. These are met, as one would expect, by changing A sufficiently often when the universal prefix-free machine U threatens to demonstrate that A has the same H-complexity as lW up to the additive constant e. We will discuss this further below. As we see below, versions of P type requirements generate followers. The R requirements refine the collections of requirements into a well-behaved stream. Their action is purely negative. If a version of P guessing that the outcome of R, is 00 (and so associated with some node p 2 a-00) generates a follower z, then that will only happen at an anco (i.e., a-expansionary) stage SO. (Note that any follower with a weaker guess will be canceled at such a stage.) Then at the next a-co stage, we will confirm the number z. This means that we cancel all numbers 2 z (which will necessarily be weaker than x). Thus x can only enter A after it needs to and a t a ,B
128
stage, at which point it must be @-confirmed. (That is, a-confirmed for all a-00 0.) Finally, we dump in the sense that if we ever enumerate x into A at stage s, then we promise to also enumerate z for x 6 z 6 s into A. It is a standard argument to show that in the limit, for any follower x that survives a-oo-stages, A 1 x can be computed from the least a-00stage s1 > SO where B,, 1 SO = B so (assuming the standard convention that uses at stage s are bounded by s). Similarly, it is also a standard argument to prove that if x is the least a-confirmed follower appointed at some a-oo-stage t with [(e, i, s) > z , then B z = B, z , where s > t is the least a-oo-stage with A , 1 x = A 1 x. More details can be found in, for example, Downey [5]. Returning to the P requirements, we build a prefix-free machine M via the Kraft-Chaitin Theorem. By the Recursion Theorem we know the coding constant d of M . We split the domain of M up into chunks for the various requirements. That is, each P, on the priority tree will be allowed to add at most 2-k((0)to the measure of the domain of M , where the sum of 2-k((0)over all strategies Pa is one. Suppose that Pa is the version of P, on the true path, and we are at stage where this version has priority (i.e., the construction never again moves to the left of P,). Let 2-k be the amount of the measure of the domain of M devoted to P,. We wait until there are a large number of a-confirmed followers for Pa (the exact number necessary is not hard to compute from k and d ) . Specifically, we pick a big x1 at an a-stage, then when the total length of agreement is bigger than 21, we initialize and a-confirm, as usual. Then we pick 22,and so on, until the whole entourage of followers is stable. Let x, be the largest of our followers. This is where we will satisfy P,. The first action is to enumerate ( k , l z n ) as an axiom for M . Thus we are saying that the H-complexity of 1"" is at most k d. If we see U(T)$= A 1 x,, with 171 6 k d e, then we change A. This is done using the followers in reverse order, first x,, then later x,-1 if necessary, and so on. The reverse order guarantees the contiguity as usual. Note that U has the option of later choosing something shorter than k d to compute l z n ,but this can only happen k d times, and we have enough xi's to cope with this. The remaining details consist of implementing this strategy on a priority tree. 0
r
r
+
+ +
+
+
129
Array noncomputable sets have one further connection with our investigations. Recall that a set A is low for random iff every random set is still random relative to A. KuEera and Terwijn [15] were the first to construct such sets. They used a theorem of Sacks [25] to prove that any low for random set A must be of GL1 Turing degree. That is, A @ 0' =T A'. This was improved by Nies [22], who also showed that there are only countably many low for random sets, and that they are all A; and hence low (i.e., A' ST 0'). The following result seems to be a theorem of Zambella. Following Ishmukhametov [13] we call a set A traceable or weakly computable if there is a computable function f such that for all g
<
(1) IWqz)I f ( z ) for almost all x and (2) g(z) E Wh(,) for all z.
Ishmukhametov [13] observed that if a degree is weakly computable then it is array computable, and the notions coincide for computably enumerable sets. Ishmukhametov proved the remarkable theorem that the computably enumerable degrees with strong minimal covers are exactly the weakly computable degrees. Furthermore, any weakly computable degree (in general) has a strong minimal cover.
Theorem 7.7. Suppose that A is low f o r random. T h e n A i s low (Nies [22]). Furthermore, A is weakly computable. Proof sketch. If one mimics the proof by Terwijn and Zambella [28] that Schnorr low sets are computably traceable, but using Martin-Lof lowness in place of Schnorr lowness, then the "if" direction proves the theorem. 0
Acknowledgments Some of the material in this paper was presented by Downey in his talk Algorithmic Randomness and Computability at the 8th Asian Logic Meeting in Chongqing, China. A preliminary version of this paper appeared as an extended abstract in Brattka, Schroder, and Weihrauch (eds.), Computability and Complexity in Analysis, Malaga, Spain, July 12-13, 2002, Electronic Notes in Theoretical Computer Science 66, vol. 1, 37-55.
References 1. Ambos-Spies, K. and A. KuEera, Randomness in computability theory, in Computability Theory and its Applications (Cholak, Lempp, Lerman, Shore,
130
2. 3.
4. 5. 6.
7. 8.
9.
10.
11.
12. 13.
14. 15. 16. 17. 18. 19.
eds.), Contemporary Mathematics 257, Amer. Math. SOC.,Providence, 2000, 1-14. Calude, C., Information Theory and Randomness, an Algorithmic Perspective, Springer-Verlag, Berlin, 1994. Chaitin, G., A theory of program size formally identical t o i n f o r m a t i o n theory, Journal of the Association for Computing Machinery 22 (1975), 329-340, reprinted in [4]. Chaitin, G., Information, Randomness & Incompleteness, 2nd edition, Series in Computer Science 8, World Scientific, River Edge, NJ, 1990. Downey, R., A; degrees and transfer theorems, Illinois J. Math 31 (1987), 419-427. Downey, R. and D. Hirschfeldt, Aspects of Complexity (Short courses in complexity from the New Zealand Mathematical Research Institute Summer 2000 meeting, Kaikoura) Walter De Gruyter, Berlin and New York, 2001. Downey, R. and D. Hirschfeldt, Algorithmic R a n d o m n e s s and Complexity, Springer-Verlag, in preparation. Downey, R, D. Hirschfeldt, and G. Laforte, R a n d o m n e s s and reducibility, in Mathematical Foundations of Computer Science 2001 (Sgall, Pultr, P. Kolman, eds.), Lecture Notes in Computer Science 2136, Springer, 2001, 316-327. Downey, R., D. Hirschfeldt, and A. Nies, Randomness, computability, and density, S I A M Journal on Computing 31 (2002) 1169-1183 (extended abstract in proceedings of STACS 2001). Downey, R., C. Jockusch, and M. Stob, Array nonrecursive sets and multiple permitting arguments, in Recursion Theory W e e k (Ambos-Spies, Muller, Sacks, eds.) Lecture Notes in Mathematics 1432, Springer-Verlag, Heidelberg, 1990, 141-174. Downey, R., C. Jockusch, and M. Stob, A r r a y nonrecursive degrees and genericity, in Computability, Enumerability, Unsolvability (Cooper, Slaman, Wainer, eds.), London Mathematical Society Lecture Notes Series 224, Cambridge University Press (1996), 93-105. Fortnow, L., Kolmogorov complexity, in [6], 73-86. Ishmukhametov, s., W e a k recursive degrees and a problem of Spector, in Recursion Theory and Complexity (Arslanov and Lempp, eds.), de Gruyter, Berlin, 1999, 81-88. Kautz, S., Degrees of R a n d o m Sets, Ph.D. Diss., Cornell University, 1991. KuEera, A. and S. Terwijn, Lowness for the class of r a n d o m sets, Journal of Symbolic Logic 64 (1999), 1396-1402. Kummer, M., Kolmogorov complexity and instance complexity of recursively enumerable sets, SIAM Journal on Computing 25 (1996), 1123-1143. Kurtz, S., Randomness and Generacity in the Degrees of Unsolvability, Ph.D. Thesis, University of Illinois at Urbana-Champaign, 1981. Ladner, R. E. and L. P. Sasso, Jr., T h e weak truth table degrees of recursively enumerable sets, Ann. Math. Logic 8 (1975), 429-448. Li, M. and P. Vitanyi, An Introduction t o Kolmogorov Complexity and its Applications, 2nd edition, Springer-Verlag, New York, 1997.
131
20. Loveland, D., A variant of the Kolmogorov concept of complexity, Information and Control 15 (1969), 510-526. 21. Martin-Lof, P., T h e definition of random sequences, Information and Control 9 (1966), 602-619. 22. Nies, A,, Low for random sets are A:, Technical Report, University of Chicago, 2002. 23. Nies, A., Lowness properties of reals and randomness, to appear. 24. Nies, A., Reals which compute little, to appear. 25. Sacks, G., Degrees of Unsoluability, Princeton University Press, 1963. 26. Soare, R., Recursively enumerable sets and degrees, Springer, Berlin, 1987. 27. Solovay, R., Draft of a paper (or series of papers) o n Chaitin’s work, unpublished notes, May, 1975, unpublished manuscript, IBM Thomas J. Watson Research Center, Yorktown Heights, NY, 215 pages. 28. Terwijn, S. and D. Zambella, Algorithmic randomness and lowness, Journal of Symbolic Logic 66 (2001), 1199-1205. 29. van Lambalgen, M., Random Sequences, Ph. D. Diss. University of Arnsterdam, 1987. 30. Vereshchagin, N., A computably enumerable undecidable set with low prefix complexity: a simplified proof, Electronic Colloquium on Computational Complexity, Revision 01 of Report TR01-083. 31. Zambella, D., O n Languages with simple initial segments, Technical Report, University of Amsterdam, 1990.
132
POPPER AND MILLER, AND INDUCTION AND DEDUCTION*
ELLERY EELLS Department of Philosophy University of Wisconsin-Madison Madison, WI 53706 U.S.A. ernail: eteellsii?facstaff.wisc.edu
In a 1983 letter t o Nature, Karl Popper and David Miller argued that there is no such thing as probabilistic inductive support. Their argument has received considerable attention, both pro and con, and in a long 1987 article in Philosophical Transactions of the Royal Society, Popper and Miller responded t o critics. And discussion continues, from time to time. Their letter, titled “A Proof of the Impossibility of Inductive Probability,” is interesting not just because of its place in the debate over the possibility of an inductive interpretation of probability, but also for what thinking about the argument can reveal about the nature of inductive evaluation of arguments. I (1988) have already joined those who find fault in Popper and Miller’s argument and disagree with their conclusion. As to the other issue, the way Popper and Miller’s argument unfolds gives a good setting for investigating the structure of an inductive evaluation of an argument (that is, the structure of a final inductive assessment of a n argument). For Popper and Miller claim that support for any given hypothesis can be divided into two components - deductive and inductive - corresponding to what they identify as two components (conjuncts) of the hypothesis the part (conjunct) that is logically implied by the evidence in question and the part (conjunct) that “goes beyond” this evidence. By thinking carefully about whether and how a hypothesis and support for it may each divide into two such natural components, I think we can in this context learn something about the nature of inductive support and some not obvious ways in which it, properly conceived, differs from deductive support. Of course deductive support is usually characterized as support in which the premises provide conclusive grounds for a conclusion where for inductive
133
support the grounds may be good but not conclusive. But there is more to it, as we shall see. 1. The Popper-Miller Argument
Popper and Miller point out that given any hypothesis h and evidence sentence e, h is logically equivalent to the conjunction ( h V e)&(hV We). Let us assume for now, as Popper and Miller do and is common, that the probabilistic support of a hypothesis h given by some evidence e is measured by the digereme, s(h,e ) = P r ( h / e )- Pr(h). (Here, P r ( z / y ) is the probability of x conditional on y, or Pr(z&y)/Pr(y).)For the purposes of their argument, nothing much hinges on the interpretation of probability, since they are arguing for the impossibility of a certain interpretation, but it is helpful to keep in mind a subjective interpretation, where P r ( z ) is a measure of an individual’s rational degree of belief in proposition or sentence, etc., x. It is easy to show that
s(h,e) = s(h V e , e )
+ s ( h V -e,
e).
Let us further assume for convenience that h and e are logically independent propositions and that h , e and their nontautologous and noncontradictory Boolean combinations all have nonextreme (not 0 and not 1) probabilities. Then Popper and Miller point out that the first addend above must be positive and the second addend must be negative. Further, Popper and Miller identify h V e as “the part” of h that is deductively implied by e and h V -e as “the part” of h that “goes beyond e.” This is because (1) h V e is the logically strongest proposition that is logically implied both by h and by e and (2) h V -e is the logically weakest proposition that is not logically implied by h and that in conjunction with h V e logically implies h. Intuitively, h V e is all of h that is logically implied by e and h V w e is the rest of h. From the facts that h can be decomposed in this way and that e probabilistically supports the part deductively implied by e and probabilistically countersupports the part that goes beyond e, Popper and Miller conclude that “All probabilistic support is purely deductive: that part of a hypothesis that is not deductively entailed by the evidence is always ... countersupported by the evidence.” (1983, p. 688) And they correctly point out that their argument “is completely general; it holds for every hypothesis h; and it holds for every evidence e, whether it supports h , is independent of h , or countersupports h.” (1983, p. 688)
134
Thus, Popper and Miller seem to see a connection between (1) dividing a hypothesis into parts and (2) dividing the kind of support a hypothesis receives into parts: the “part” of h that is deductively implied by e receives just the deductive kind of support and the part of h that is not deductively implied by e receives only the inductive (or other, or nondeductive) kind of support. And the former is always positive, probabilistically speaking, and the latter is always negative, probabilistically speaking. As I see it, it is this supposed connection between components of a hypothesis and kinds of support the hypothesis receives from the evidence that is supposed to justify the conclusion that, probabilistically speaking, only the deductive kind of support is positive and the inductive (or the nondeductive) kind of support is always negative (%ountersupport”). There are thus, as I see it, two main issues here that deserve attention, and which are important not only for the evaluation of the Popper-Miller argument but also for a proper understanding of the nature of and the difference between inductive and deductive support. The first concerns the way in which Popper and Miller have divided a hypothesis into parts, Other ways have been suggested as more or less or equally appropriate; some of these other ways will be discussed below. The second issue has to do with the idea of dividing the support a hypothesis receives into two components or kinds in a way parallel to the different kinds of support relations evidence supposedly bears to parts of a hypothesis. I will argue that Popper and Miller have incorrectly (or, more specifically, not fully) assessed the kind of support h V e receives from e and that it is incorrect to think of the support relation between premises (or evidence) and conclusion (or hypothesis) as even being either of only the deductive kind or only of the inductive (or other, nondeductive) kind. I will suggest (an idea that is not new) that an evidence/hypothesis (or premises/conclusion) pair - that is, an argument - is not itself either inductive or deductive, but rather the difference between induction and deduction (between inductive logic and deductive logic) has rather to do with different ways or standards imposed for the evaluation of any fixed such pair. This will suggest a clarification of the form of deductive and inductive evaluations (that is, final assessments) of arguments (evidence/hypothesis or premises/conclusion pairs).
2. “Parts” of h
As I mentioned, Popper and Miller have been challenged in connection with the way they divide a hypothesis h into parts. One challenge, by
135
Richard Jeffrey (1984) which has been widely discussed, focuses on the part of h that Popper and Miller characterize as the part of h that “goes beyond e.” Jeffrey considers hypotheses that are universalized conditionals (“If anything is a so-and-so, then it is a such-and-such”, or “All so-and-so’s are su”-and-~uch’~’’).Let h be “All emeralds are green” and let e be “The many emeralds we have observed have all been green”. Jeffrey says that the part of h that goes beyond e is clearly f : “SOare the rest (that is, all unobserved emeralds are green)”. On Popper and Miller’s formula the part of h that goes beyond e would be “Either all emeralds are green or it is not the case that we have observed many emeralds, all green”. Certainly f seems, at first sight, more natural than Popper and Miller’s h V -e as a formulation of the part of h that goes beyond e , in this case. Furthermore, on this f -understanding, it is also natural to think that e positiveIy supports the part of h that goes beyond e , that is, that the proposition ( e ) that many emeralds, all green, have been observed positively probabilistically supports the hypothesis (f) that the unobserved emeralds also are green (at least this is true of my own conditional and unconditional subjective probabilities, or degrees of belief). And this is contrary to the probabilistic significance of e for h V -e - thus undermining the Popper-Miller claim that evidence can only probabilistically countersupport the part of h that goes beyond e. However, as Jeffrey points out, it is not in general true on the f -understanding, that evidence probabilistically supports the part of h that goes beyond e. For example, let h’ be “All emeralds are grue” (where “grue” means “observed and green, or unobserved and blue”) and let e‘ be the corresponding evidence. Then f’, the part of h that goes beyond h’ on Jeffrey’s understanding of this, is equivalent to “All unobserved emeralds are blue”. While it is plausible that e probabilistically positively supports f it is also plausible that e’ probabilistically countersupports f‘. And Jeffrey claims that h V -e (and h‘ V -el) “is another matter altogether . . . [Tlhe relevant factoring is . . . h = f &e, where f is no truth function of h and e.” (1984, p. 433) There is a kind of paradox here, in that, when read separately, both Popper and Miller’s and Jeffrey’s formulations of the part of h that goes beyond e seem plausible [to me anyway). I think this little paradox can be diagnosed and resolved by looking at the different contexts within which Popper and Miller, on the one hand, and Jeffrey, on the other, formulate their “ampliative” (or “going beyond the evidence”) parts of a hypothesis. First, Popper and Miller’s discussion is not in the context of any concrete example (their 1983 letter in fact contains no examples a t all), while Jef-
136
frey’s f -suggestion is in the context of hypotheses that are of the particular form of universalized conditionals, and he discusses two particular examples, the green and grue hypotheses. Jeffrey’s approach would seem to work when the hypothesis is a universalized conditional, but it is not at all clear how it could be applied to hypotheses of other logical forms, such as existential, disjunctive, or singular propositions. Popper and Miller’s h V -e would seem to apply equally well to all evidence/hypothesis pairs (whether this is very well or not so well is a separate question); and recall that POPper and Miller’s argument is “completely general,” as they say. Second, it is instructive to look at just why Jeffrey’s f-approach seems natural in the context in which he makes his suggestion. The hypothesis that all emeralds are green can be thought of as saying something about all things in a certain ”domain,” the class of emeralds. The hypothesis says that all things in that class are green. (At least the class of emeralds is one way of thinking of what the hypothesis is about. It also says of every thing that it is either green or not an emerald. But this point is consistent with the point I wish to make below.) The class of emeralds can, in a completely straightforward way, be divided into two parts: the observed ones and the unobserved ones. The evidence e covers the first part of what h is about, and f covers the second - the rest, the part “beyond e.” That is why h is logically equivalent to f &e, the two conjuncts separately covering the disjoint and exhaustive parts of the domain that h is about. What about Popper and Miller’s h V e and h V Ne? Can we identify, in a general way, the classes of things that h, h V e, and h V -e are about? Maybe, but the relevant classes would be of quite different kinds of things from what they are in the natural way of understanding what they are in the special case of universalized conditionals. It is perhaps helpful to think in terms of Venn diagrams. In Jeffrey’s example we can draw a large box for the universe of discourse and three overlapping circles corresponding to emeralds, green things, and observed things, for eight disjoint regions inside the box. The points in the circle represent possible things in the actual world, and shading a region represents the proposition that there are no actual things corresponding to points in that region; we can thus represent the hypotheses e , h, and f . We cannot, however, draw a diagram that represents the propositions h, h V e, and h V -e in general (that is, for arbitrary h and e ) and interpret the points in the region in the same way. Of course, we can draw circles representing h and e and indicate combinations of regions that represent propositions h V e , h v we, and so
137
on, but the interpretation is different. Here, the points would represent possible worlds, or possible situations, or fully specific logical possibilities, and we cannot say that h is about things that are represented by points in a certain region. A single diagram of the first kind represents relations among classes of things in a given possible world (the world the relevant propositions are about), while a diagram of these cond kind represents relations among propositions - which can be thought of as classes of possible worlds (represented by regions in the diagram). Thus, one way of diagnosing the “conflict” between Popper and Miller’s and Jeffrey’s h V me- and f-ideas, respectively, is to say that two different ideas of “parts” of a hypothesis are operating: while Jeffrey divides a hypothesis up according to what in the world the hypothesis is about, Popper and Miller divide a hypothesis up according to what other propositions, classes of worlds, imply or are implied by the hypothesis. For Jeffrey, “parts” corresponds to subclasses of a universe of discourse of objects in a given world; for Popper and Miller, LLparts” corresponds to classes of possible worlds that stand in certain logical relations to the hypothesis. To put it yet another way, Jeffrey focuses on parts of the actual world, and Popper and Miller focus on logical parts of a hypothesis (what things it says, or logically implies). From this point of view, and since the Popper-Miller argument works for hypotheses in general (not just for universalized conditionals) if it works at all, it would seem that Jeffrey’s f-formulation of the part of h that goes beyond e is not to the logical point. In saying this, I do not mean to imply that Jeffrey’s f -formulation does not represent a natural understanding of the idea of “going beyond the evidence,” even more natural than Popper and Miller’s in the case of universalized conditionals. The point is just that the f-understanding does not come to terms with the way in which Popper and Miller factor a hypothesis, a logically coherent way that seems worth evaluating on its own terms. On the other hand, there certainly are logical relations among e , f , and h that have interesting bearings on probabilistic support relations among them. While I may be alone in thinking that Jeffrey’s f-part idea is off point when thinking about the parts of a hypothesis in the context of the Popper-Miller argument, a close look at some discussions, by other authors, of the relations among e , f , and h will prove instructive later, when we turn to the different structures of deductive and inductive evaluations of arguments. (Barman (1992, p. 98), like Jeffrey, also discusses confirmation of universal generalizations in the context of the Popper-Miller argument .)
138
Another interesting challenge to Popper and Miller’s way of dividing a hypothesis into parts was raised by Michael Dunn and Geoffrey Hellman (1986). While Popper and Miller divide a hypothesis into two disjunctive conjuncts, Dunn and Hellman point our that another way to “factor” h is into conjunctive disjuncts. That is, h is logically equivalent to (h&e)V (h&-e). Since this is the result of interchanging “V” and “&” in Popper and Miller’s factorization, Dunn and Hellman call this the “dual” of Popper and Miller’s factorization. Furthermore, it is easy to show that
s(h,e ) = s(h&e,e)
+ s(h&-e,e),
where the first addend is necessarily positive and the second necessarily negative (given our simplifying assumption of logical independence of h and e and of nonextreme probabilities for nontautologous and noncontradictory Boolean combinations of h and e ) . And since e does not logically imply h&e, while e does logically imply -(h&-e), Dunn and Hellman conclude (tongue-in-cheek, I think) “that all countersupport is ‘purely deductive ’. Any(positive) support must be contributed by the first component, which clearly is not purely deductive. So (contra Popper and Miller) not all support is purely deductive.” (1986, p. 222) I think the correct reading of Dunn and Hellman (1986) is not that factoring into a disjunction of conjunctions is right and the conjunction of disjunctions is the wrong way to factor, but rather that there is nothing about the two ways that makes one better than the other for the purposes at hand. (This reading, by the way, is contrary to Colin Howson’s (1989) reading, who reads Dunn and Hellman as putting forth their dual as a superior factorization, while in fact all Dunn and Hellman say is more or less that it’s a matter of taste and that they happen to prefer thinking in terms of their dual representation.) I agree with Dunn and Hellman’s indifference between the two factorizations; I think this indifference can be justified; and I think that this alone is devastating to the Popper-Miller argument (though there is more to say in the way of diagnosing the failure of the argument and something to learn from the diagnosis). Of course, given any partition {el, . , en} of propositions, a hypothesis h can be expressed in these different logically equivalent ways: 1
h = A h V ei, and h = i
v
h&ei.
i
And this is not just an abstruse logical fact that has no relevance to the way we often do, or how we properly should, think of the parts of a hypothesis.
139
Suppose I wanted to tell a student that a certain theorem is stated somewhere between pages 16 and 19 of the text (inclusive). I could tell her: “It’s on page 16 or 17 or 18 or 19.” Or I could tell her: “It’s after page 15 and before page 20.” While “between 16 and 19” may be the most natural way of expressing what I remember about where the theorem is stated in the text, this illustrates the simple idea that there are at least these two ways of expressing exactly what a proposition (or hypothesis h) says: You can either enumerate everything that has to be true for the proposition to be true or enumerate everything that would make the proposition true. That is, one can enumerate either all the necessary conditions or all the sufficient conditions. The difference between Popper and Miller’s factorization and Dunn and Hellman’s factorization is simply that Popper and Miller conjoin two necessary conditions for h that are in conjunction sufficient for h , while Dunn and Hellman disjoin two sufficient conditions for h that are in disjunction necessary for h. Of course, different ways of carving up and enumerating all the necessary or all the sufficient conditions may be more or less natural or easy to grasp in different contexts, and in some cases not both of these ways may be practically or theoretically available. But it is easy to multiply examples in which each of the duals is equally “natural.” It is worth mentioning that some authors have explicitly argued in favor of the Popper- Miller factorization. The issue involves the proper formalization of the idea of “content” (or “logical content”) of a proposition. Popper and Miller (1987, p. 580) identify the content of a proposition IC with its consequence class, Cn(z),the set of all the logical consequences of 2. And in an endnote addressing Dunn and Hellman’s “dualling,”they state, “It appears to be their view that there is nothing to choose between [their dual factorization and conclusion] and ours, and so neither can be correct. But unlike h V e and h V -e, neither h&e nor h&-e is a part of the content of h.”(p. 590) The idea is that h&e and h&-e are not, properly conceived, parts of h since they are not logical consequences of h, while hVe and hV-e are. Similarly, Colin Howson (1989) says, “it seems to me that where a and b are statements a necessary condition for b’s representing a proper part of a’s content is that b is a strict, or one-way, consequence of a. This is a pretty minimal condition, but I think that if anything is an analytic truth then that is ....However, h&e and h&-e are each stronger than h if h and e are independent and therefore even on this minimal condition cannot represent proper parts of h.” (1989, p. 667) Howson follows with a discussion of inclusion relations between Cn(a) and Cn(b),between M ( a ) and M ( b ) (where
140
M ( z ) is the class of models of z), and between M(-a) and M(-b), when a logically implies b and for other relations among propositions. He notes that M ( h ) = M(h&e)U M(h&-e), a disjoint union, and explains that this is the reason for Dunn and Hellman’s preference for their dual factorization of h into a disjunction of mutually exclusive conjunctions (but recall from above that I don’t think it’s a real preference on Dunn and Hellman’s part). However, he complains that M does not get the set theoretical inclusion order right for content: Cn(a) 5 Cn(b) if and only if M ( b ) M ( a ) . And he suggests that “in the context of a discussion of content there would be more reason to identify [u] with [M(-a)] rather than with M ( a ) : only with the former does one get an ordering by inclusion which agrees with that by inclusion between consequence classes.” (p. 667) My assessment of these reasons for preferring Popper and Miller’s decomposition is that it just reflects a preference for a full enumeration of necessary, rather than of sufficient, conditions in expressing what a proposition says, where in fact there is no principled basis for preferring one over the other in general. To put it another way, in the context of this set theoretical discussion, to express what h says one could either describe the classes of models of the necessary conditions for h (and take the intersection of these classes) or describe the classes of models of the sufficient conditions for h (and take the union of these classes). That is, the Popper-Miller factorization corresponds to the fact that M ( h ) = M ( h V e ) n M ( h V -e) and the Dunn and Hellman factorization corresponds to the fact that M ( h ) = M(h&e) U M(h&-e). And, of course, analogous to the two syntactical equivalences displayed above (also when prefixed by “Cn”),here we have, where {ei}i is any partition,
M(h)=
n
u
i
i
M ( h V e i ) , and M ( h ) =
M(h&ei).
Further, I do not see the force of the objection to taking M as a measure of content that says that M gets the inclusion order of content wrong; A4 is simply a dual measure to Cn, as Cn(a) Cn(b)if and only if M ( b ) C M ( a ) . The objection seems a lot like objecting to the scoring system in golf by saying that it always gets the rankings exactly reversed. Because of the dualities and symmetries discussed above, I conclude that there is nothing to favor the Popper-Miller factorization and conclusion over Dunn and Hellman’s factorization and “conclusion,” and vice versa. Since the two conclusions contradict each other, this is a reductio of both conclusions (which I take to be Dunn and Hellman’s point). Later, I turn
141
to diagnosis.
3. “Components” of Support
+
As we have seen, in the equation s(h,e ) = s(h V e , e ) s(h V -e, e), Popper and Miller identify s(h V e , e ) as the deductive component of e’s support for h. and s(h V -e,e) as the inductive component of this support. For now, let’s focus on the first addend in the equation. In their (1987) paper, Popper and Miller speak of degrees of deductive dependence, and they begin their article saying that they shall interpret probability as “logical probability,” without explaining just how to interpret probability as “logical probability.” But they say that the particular axiomatization isn’t crucial to their argument and that they intend their argument to be completely general. In any case, it seems that we should distinguish between “deductive support” and ‘‘support for a hypothesis that is deductively implied by the evidence.” In s ( h V e , e ) we at least have the second thing. The question now is whether there is some interesting sense, beyond this, in which we can say that s ( h V e , e ) measures support that is purely deductive in nature. It seems to me clear that the answer is no: s ( h V e , e ) obviously carries more information, and in a way less, than just whether or not the evidence ( e ) deductively supports the hypothesis ( h V e in this case), in the sense of deductively implying the hypothesis. For s ( h V e , e ) is a difference, in this case where the evidence logically implies the hypothesis: 1 - P r ( h V e ) . Thus s ( h V e, e ) depends on the value of the “prior probability” Pr(h V e ) , a value that is of course completely independent of the deductive relations between h V e and e. This already strongly suggests that both addends in the equation above represent nondeductive (or inductive) relevance to parts of h - s ( h V e , e ) representing nondeductive positive support for h V e by e (or at least not purely deductive support since it is a function of Pr(hV e ) ) and s ( h v We, e ) representing nondeductive countersupport for h V -e (as Popper and Miller agree). What this suggests is that overall probabilistic support of h by e is a sum of (i) inductive positive support for a part deductively implied by e and (ii) inductive counter support for a part of h not deductively implied by e. In (i) we have positive inductive support of a deductively implication, and in (ii) we have negative inductive support for a deductive nonimplication. Even from this, however, we should not conclude that the overall probabilistic support for h by e is not positive inductive support for a hypothesis that is not deductively implied by e. As Charles Chihara has pointed out (see
142
Chihara and Gillies 1988)we should not conclude from the fact that neither of two functions represent some quality that their sum does not. In his example, he lets h ( z ) = (1/2)p(z) 2 and j ( z ) = (1/2)p(z) - 2, where p is a probability function. Neither h nor j is a probability function, but their sum, p , is. Chihara actually applies his example in a slightly different way. Even if we granted that (i) above should be regarded as “purely deductive support,” it does not follow that the sum of purely deductive support and inductive countersupport cannot represent inductive support. (See Gillies and Chihara (1988) for a reply by Gillies.) However, I do not wish to grant that s(h v e , e ) represents purely deductive support. If the fact that this value is a function of P r ( h V e ) is not already fully convincing, consider the following pair of probabilistic/logical examples (reproduced from Eells (1988)). Here, e / h and E / H are two evidence/hypothesis pairs (alternatively, the two examples could be expressed using a single evidence hypothesis pair e/h and two probability functions, P r and PR, which may represent rational subjective probabilities of different people).
+
EXAMPLE 1:
Pr(h&e) = .3 Pr(h&-e) = .2
Pr(-h&e) = .2 Pr(-h&-e) = .3
so: P r ( h / e ) = .6 Pr(h v e/e) = 1 Pr(h V -e/e) = .6
P r ( h ) = .5 Pr(h v e ) = .7 P r ( h V -e) = .8
s ( h , e ) = .I s(h V e,e) = .3 s(h V -e, e ) = -.2
EXAMPLE 2:
Pr(H&E) = .3 P r ( H & - E ) = .4
Pr(-H&E) = .2 Pr(-H&-E)
= .1
so:
P r ( H / E ) = .6 Pr(H V E/E)= 1 P T ( H V - E / E ) = .6
P T ( H ) = .7 P T ( H V E ) = .9 P T ( H V W E )= .8
s ( H , E ) = -.l s ( H v E , E ) = .1 s ( H V W E ,E ) = -.2
Across the two examples, the deductive relations among the propositions involved are exactly parallel, and the probabilistic inductive countersupports are the same, as s(hV-e,e) = s ( H V - E , E ) = -.2. Since all the deductive relations are the same in the two examples and since the overall support for the two hypotheses h and H are different (.l and -.l),there must be
143
a nondeductive difference between the two examples. This nondeductive difference is not in the respective nondeductive countersupports (which is -.2 in both cases). So the nondeductive difference must be a difference in probabilistic support features in the two examples, that is, in s(hVe,e) and s(H V E , E ) . It follows that s ( h V e, e) and s ( H V E , E ) do not represent purely deductive aspects of the evidences’ support of the hypotheses. It is perhaps worth noting that pairs of examples like this can be constructed also for Dunn and Hellman’s dual factorization of the support for h by e, s(h,e ) = s(h&e,e) + s(h&-e, e ) , with the “dual lesson” that the countersupport of h&-e provided by e should not be thought of as purely deductive countersupport. But I will not bother the reader with such a pair of examples here. I just note that though Pr(h&-e/e) = 0, s(h&-e,e) is a function of the prior probability Pr(h&-e), which shows that the countersupport for h&-e provided by e is not purely deductive. Thus, I think the correct way of viewing the Popper-Miller decomposition of support (and the Dunn and Hellman dual decomposition of support) is to think of the inductive support of h by e as decomposable into positive inductive support of a part h v e of h that is logically implied by e and inductive countersupport of a part h V -e that is not logically implied by e (and as decomposable into inductive positive support of a part h&e of h that is not logically implied by e and inductive countersupport of a part h&-e of h whose negation is deductively implied by e ) . As to the two decompositions, we have positive inductive support for h V e and for h&e and negative inductive support for h V -e and for h&-e. In the case of the Popper-Miller decomposition, overall inductive support lies between support for h V e and for h v w e , which supports fall in ( 0 , l ) and (-1,O) respectively. And in the case of Dunn and Hellman’s decomposition, overall inductive support for h falls between support for h&e and support for h&-e, which supports again are in ( 0 , l ) and (-1,O) respectively. No doubt other interesting ways of decomposing probabilistic inductive support for h by e can be described, but when it is decomposed using e it is not surprising that a component of h surfaces that has deductive implication relations with e.
4. Inductive Assessments
It seems to me that the most natural way of thinking of deductive support is as an all- or-nothing relation: either premises x give an absolute guarantee of truth t o a conclusion y , or not. In the case where x does logically imply y, x settles the truth of y, so it is perhaps easy to overlook the
144
prior probability Pr(y) involved in s(y, x). In any case, probabilistically speaking, deductive implication is reflected just in the final, or posterior, probability Pr(y/z) = 1. In discussing inductive support, Isaac Levi (1984, 1986) endorses a condition that suggests a way of assessing inductive support analogously to the way deductive support is assessed. He puts forth the following equivalence condition for inductive support: h‘, then the inductive support of h provided by e equals the inductive support of h‘ provided by e.
ECL: If e I- h
The support function, s, that we have been discussing does not satisfy this condition, for s involves the “change” of the probability of h, from unconditional to conditional-on-e, which depends on the initial, unconditional probability, while any support function that satisfies ECL ignores this initial probability. According to ECL, as long as two hypotheses are equivalent in the presence of e, what their initial probabilities were does not matter. An alternative support function that satisfies ECL is the one that identifies inductive support with the “final,” conditional probability Pr(h/e). Following Gaifman (1985), who calls these two support ideas “suppmt,” and “suppmtj” ( “ c ” for “change” and “f” for “final”), let us say that s,(h,e) = Pr(h/e) - Pr(h) (as before) and sj(h,e) = Pr(h/e). Unlike sc, whenever e deductively implies h, s f ( h ,e) = 1. Since s, fails to satisfy ECL, Levi concludes that at least s,, which he calls “probabilistic support”, is not a measure of inductive support. Thus, as Levi puts it, he reaches Popper and Miller’s conclusion in a different way. (I should note that Levi rejects both sc and s j as measures of inductive support.) The measure sj makes deductive and inductive assessments of premises/conclusion pairs, arguments, structurally analogous in at least this way: an argument e/h is deductively valid only if (but not always if) sj(h,e) = 1, and the inductive strength of an argument is measured by whatever value sj(h, e) has. Roughly, then, deductive logic is concerned just with whether sj(h,e) is equal to 1 or not, whereas inductive logic is concerned with explicating sj(h, e) (that is, Pr(h, e ) ) whatever this value may be. Intuitively, we can say that deductive logic is concerned with whether or not an argument is “perfect” (premises either give an absolute guarantee of truth to the conclusion or they don’t give an absolute guarantee of truth to the conclusion), while inductive logic is concerned both with this and with gradations of imperfection. But on this approach to inductive assessments of arguments, inductive and deductive logic are nevertheless analogous in that they would both focus just on the final degree of
145
support a conclusion achieves in the presence of the premises, where change is not a factor. A further way of making inductive and deductive assessments analogous is this. In deductive logic an argument is said either to be valid or invalid, a two-possible-value assessment, whereas an inductive assessment by sf (h,e ) can take on a continuum of values. But two-valued inductive assessments have been suggested. Levi (1986) calls them “satisficing” criteria and Mary Hesse (1974) has called them “k-criteria”. Thus we can say that an argument is strong if s j ( h ,e ) > k and otherwise weak (for some chosen value of k , where it is usually suggested that k >. .5). In this way, loosely speaking, both deductive and inductive logic can be said to focus on some threshold value, 1 or some k. Of course, we should grant that all this is certainly one coherent and clear conception of inductive support (leaving aside, as we have been, the question of how probability should be interpreted). However, I think it fails to capture, or conflicts with, some ideas that we should naturally include in a formalization of the idea of nondeductive support. In deductive assessments of arguments, all we can say, probabilistically speaking, is whether or not the premises leave the conclusion at probability 1. The premises, if true, either settle the question of the truth of the conclusion or they do not, and this often answers all the questions we may have about the support the premises provide the conclusion, where in the case of a valid argument it makes no difference whether the final probability of 1 involves no increase in probability a t all, or a probability difference of .1, .3, .8, or whatever. However, in the case of inductive inference, the especially interesting cases are where the final probability is not 1 (or 0). An increase in probability by .3 could be an increase from .7 to 1, from .4to .7, from .1 to .4, and so on. In addition, perhaps an increase from .7 to 1 would be less impressive than (or at least have a different assessment from) an increase from .2 or .3 to 1. But s j ignores these latter differences involving different amounts of “boost” t o get t o probability l(or t o any other value). Also, sc, the difference measure, alone, registers just the amount of change, and n o t initial and final probabilities, which would seem to be incomplete if we count, for example, an increase from .7 to 1 as importantly different from an increase from, for example, .1 to .4. Suppose, for now, that we can agree that probability change (the difference between initial and final values) is important in an inductive assessment of an argument. Given this, we should agree further that, for example, whether a change of magnitude of .3 is from .7 to 1 or from .1
146
to .4is also important information that should be included in an inductive assessment of an argument. This seems right since an increase from .7 to 1 would settle the question of the truth of the conclusion (at least with maximal probability) while an increase from .I to .4 would not. And changes of magnitude .3 from other initial to final values would also seem to yield intuitively different assessments of arguments (for example, sometimes a threshold of “beyond a reasonable doubt” may be important). Thus, if we could agree that change is important, then we should agree that a full inductive assessment of an argument should include also the initial and final probabilities of the conclusion. Let us now return to the question of the importance of change - basically the s, or s f question. This should involve an evaluation of Levi’s equivalence condition, which I have called ECL. Recall Jeffrey’s example involving evidence e (that all the observed emeralds are green) and hypotheses h (that all emeralds are green) and f (that the unobserved emeralds are all green). On the assumption that e is true, h and f , as well as h V w e , are equivalent, so by ECL e should provide the same support for each of h, f , and h V w e . However, the measure sc disagrees with this verdict, since, as we have seen, s,(h, e ) and s,(f, e) should both be positive, while s,(h V w e , e ) is negative. For this reason Levi rejects s, as a measure of inductive support. However, s f ( h ,e), sf(f,e) and s f ( h V N e , e) are all equal. One way of seeing this disagreement is to note that h V -e should, prior to the evidence, start out at a high value: prior to observing the color of (the “observed” or to-be-observed) emeralds; the proposition that all the observed emeralds will have same color, and indeed green of all the colors there are, would seem pretty unlikely. My own inclination is to say that the evidence countersupports h V -e and supports h and f - contrary to ECL and s f which say that the three hypotheses are equally confirmed by the evidence since the three are equivalent given the evidence. However, other, simpler, examples may help to reinforce this intuition. Suppose scientists have just invented a new compound, call it substance X , but have not yet produced or examined samples of X. Suppose it is known, or just believed to degree 1,that either all samples of X will be blue or they will all be white (or just that they will all have the same color). This might be somehow based on the known chemical composition of X ; they just don’t know which color all the X’s will be. Call this hypothesis h’. Let h be the hypothesis that all X’s will be blue. Finally, let evidence e report the observation, finally, of some samples of X that they are blue. e might report multiple observations of X’s, where at first the color wasn’t certain
147
because of lighting conditions. In this case it seems clear that e gives no support at all to the hypothesis h', since it was already known that (what h' says) all X ' s would be of the same color - it was just not known which color. However, the repeated observation of X ' s under different lighting conditions - that is, evidence e - clearly lends considerable support t o the hypothesis h that all X ' s are blue. Thus, e supports h but not h'. However, given el h and h' are equivalent (given the way we understand color), so that this result conflicts with E C L . Of course, the reason we think of e as supporting h is the boost in probability that e gave to h, from some value less than 1 t o 1, while we say that e doesn't support h' because we know h' already, its probability of course remaining at 1 after e was learned. These judgements conflict also with the thesis that sf measures inductive support, and is consistent with the thesis that s, measures inductive support. For another simple example, which does not involve the probability value 1, suppose that a regular six-sided gaming die is about to be rolled. The hypotheses are h , that the side with two dots on it will come up, and h', that a side with more than one dot on it will come up. Now the die is rolled and we get a glimpse of it and can see that the up side has either one or two dots on it (or someone tells us this). This is evidence e. And suppose that we now assign equal probability to the die's having one dot up and having two dots up. It seems clear that e supports h and countersupports h'. This verdict conflicts with ECL since again h and h' are equivalent given e. It also conflicts with the thesis that s f measures inductive support and is consistent with the thesis that s, measures inductive support, since the relevant probabilities are Pr(h') = 5/6 > P r ( h / e ) = Pr(h'/e) = 1/2 > Pr(h) = 1/6. From the examples above it is easy to see that there is also this other feature shared by all four of ECL, sf,k-criteria, and the satisficing idea. They are all consistent with the possibility of strong inductive support despite the fact that the evidence decreases the probability of the hypothesis in question - in the case of ECL, this will happen when the hypothesis is equivalent given the evidence to a hypothesis that is strongly supported by the evidence. In the cases of s f ,k-criteria, and the satisficing idea, the final probability can still be high or above a threshold level even if the evidence lowered this probability from an even higher value. I conclude that each of ECL,sf,k-criteria, and the satisficing idea miss important features of an adequate inductive assessment of evidence/hypothesis pairs. Also, s, fails to register an important factor in the
148
inductive assessment of arguments. Let me recap the reasons why, focusing for simplicity just on sc and s j . The reasons can be seen in comparing what we expect from deductive and inductive assessments. In deductive assessments, all we can expect, probabilistically speaking, is the information of whether the final probability of the hypothesis is 1 or not. This misses two important pieces of information that are relevant to inductive assessments. First, inductive assessments have to take account of the possibility that the final probability is less than 1; that is why P r ( h / e ) (or sj), whatever it is, and not just whether or not it is 1, should be included. Second, we have seen that probability change - I would include both direction and magnitude of change - is relevant; that is why s, is important, whether the final probability is 1 or not. Deductive assessment is concerned only with whether or not the final probability is 1 (probabilistically speaking); inductive logic should deal in addition both with final probabilities less than 1 and with direction and magnitude of change. Finally, we have seen that another important factor in inductive assessment is just what the initial and final probabilities are; that is why sc alone is not enough, and this is made vivid in that there is an important, settling the truth of the conclusion, issue involved when the final probability is 1. The natural conclusion of all this is that an inductive assessment of an argument, or evidence/hypothesis pair, should include (at least) the values Pr(h) and Pr(h/e). Note that the conclusions above have to do with the structures of final assessments of arguments, and not with techniques of argument evaluation. In deductive logic, the assessment is either “valid” or “invalid”, while in inductive logic, the assessment is richer, I have urged, including a specification of both initial and final probabilities. In deductive logic, evaluation involves the well known idea of deductive calculi. In inductive logic the problem of evaluation is not as well developed. First, there is the problem of interpreting probability, on which I have not touched. This would seem to correspond to the semantical side of a system of deductive logic where the meanings of validity and related concepts such as consistency and logical truth are formalized in one way or another. And second, with the richer assessment “output” of evaluation that I have suggested, it would seem that the “input” would also have to be richer. In deductive logic, the input is just the assumption of the truth of the premises. In inductive logic the input would seem to have to involve not only this but also certain background knowledge or beliefs within the context of which initial and final probabilities can be evaluated. This of course is complex and perhaps intractable (I don’t mean in the technical complexity theoretic sense here).
149
Some have taken this as a fact that makes for an advantage of a subjective interpretation of probability over logical interpretations, since a person’s rational subjective probability assignment can be thought of as a kind of systematic summary of the relevant background information. It is also worth mentioning that some have suggested collapsing the two-valued inductive assessment I have urged into a one-valued measure, We have already discussed the difference measure. The quotient P r ( h / e ) / P r ( h )has also been suggested as have other, more complex, functions of P r ( h / e ) and P r ( h ) (or of P r ( h / e ) and Pr(h/-e)). I doubt that any single such measure would do as well as all the others in every context. We have already seen this for the difference measure. In some contexts a very high quotient may be impressive. For example, if P r ( h / e ) = .001 and Pr(h) = .000001, then the quotient of 1000 may be impressive from a diagnostic or epidemiological point of view, but not in some decision theoretical contexts or in a context where ”belief beyond a reasonable doubt” is important. I suggest, though, that it is no accident that all such one-valued measures (at least that I am aware of) are functions of just P r ( h / e ) and P r ( h ) (or of P r ( h / e ) and Pr(h/-e)), so that it seems that the two-valued assessment would be the most versatile, adaptable to the various pragmatic and theoretical problems in which inductive assessment of the bearing of evidence on hypothesis is pertinent. (I am ignoring here likelihood measures, which involve probabilities of evidence conditional on hypotheses that make different predictions; such measures may be helpful in assessing differential support for such competing hypotheses.)
5 . Conclusion
Returning to the Popper-Miller argument, I would say that the problems we have seen with it stem from at least one of two possible sources. First, there may have been a conflation between the ideas of deductive support and support of a hypothesis that is deductively implied b y the evidence. As I mentioned earlier, in s,(h V e , e ) , we certainly have the latter, since e deductively implies h V e. But it is perhaps misleading to count this as only deductive support, or “purely deductive support.” For it seems to me that any argument can be evaluated both in terms of deductive support and in terms of inductive support. There are not two distinct kinds of argument, inductive arguments and deductive arguments (though it might be clear in context that the author of an argument intends the argument to be evaluated “on his or her own terms,” inductive or deductive). Rather, any given
150
fixed argument can be evaluated both deductively and inductively. When an argument is deductively valid, deductive logic can tell us that the conclusion is necessary given the premises (which information inductive logic may leave out, since probability 1 is not necessity), but gives no probability change information, which a fully adequate inductive logic would give. When an argument is not deductively valid, deductive logic can tell us that the conclusion is not necessary given the premises, but gives us no information about the (inductive) impact the premises have on the conclusion, which a fully adequate inductive logic would give. Inductive and deductive logic just evaluate the same arguments on different terms, where there is no such thing as “purely deductive support” or “purely inductive support.” While the first possible source of problems with the Popper-Miller argument may derive ultimately from a perhaps unfortunate choice of terminology (which I think it is nevertheless worth becoming clear about), the second possible source is more substantial. This is simply the difference between different conceptions of inductive support discussed above. If we accept the view that inductive support involves not only final probabilities - the value of 1 most relevantly in the Popper-Miller argument and 0 most relevantly in the Dunn and Hellman dual formulation - but also a n incremental component, then it becomes clear that neither the Popper-Miller decomposition nor the Dunn and Hellman dual decomposition involves a decomposition into purely deductive support (or countersupport) and purely inductive countersupport (or positive support). In the case of each factorization, we have a decomposition of a hypothesis into two components, where the support of each of the two components can be evaluated and assessed, nontrivially, both inductively and deductively. While I think the Popper-Miller argument is fallacious in concluding that there is no such thing as probabilistic inductive support, I think it provides a fruitful context for investigating the difference between deductive and inductive assessments of arguments, and the nature of the latter.
* I thank Mike Byrd for helpful comments on an earlier draft of this paper.
REFERENCES Chihara, C. S. and Gillies, D. A. (1988), “An Interchange on the PopperMiller Argument,” Philosophical Studies 54: 1-8. Dunn, J. M. and Hellman, G. (1986), “Dualling: A Critique of an Argument of Popper and Miller,” British Journal for the Philosophy of Science 37: 220-223.
151
Earman, J . (1992), Bayes or Bust?: A Critical Examination of Bayesian Confirmation Theory,The MIT Press, Cambridge, Massachusetts. Eells, E. (1988), “On the Alleged Impossibility of Inductive Probability,” British Journal forche Philosophy of Science 39 111-116. Gaifman, H. (1985), “On Inductive Support and Some Recent Tricks,” Erkenntnis 2 2 5-21. Hesse, M. B. (1974), The Structure of Scientific Inference, University of California Press,Berkeley and Los Angeles. Howson, C. (1989), “On a Recent Objection to Popper and Miller’s ‘Disproof’ of Probabilistic Induction,’’ Philosophy of Science 56: 675-680. Jeffrey, R. C. (1984) , “The Impossibility of Inductive Probability,” Nature 310: 433. Levi, I. (1984), “The Impossibility of Inductive Probability,” Nature 310: 433. Levi, I. (1986), “Probabilistic Pettifoggery,” Erkenntnis 2 5 133-140. Popper, K. R. and Miller, D. W. (1983), “A Proof of the Impossibility of Inductive Probability,” Nature 302 687-688. Popper, K. R. and Miller, D. W. (1987), “Why Probabilistic Support is not Inductive,” Philosophical Transactions of the Royal Society London A 321: 569-591.
152
ENLARGEMENTS OF POLYNOMIAL COALGEBRAS*
ROBERT GOLDBLATT Centre f o r Logic, Language and Computation
Victoria University of Wellington, New Zealand E-mail: Rob.Goldblatt@vuw. ac.nz We continue a programme of study of the model theory of coalgebras of polynomial functors on the category of sets. Each such coalgebra a is shown to have an “enlargement” to a new coalgebra Ea whose states are certain ultrafilters on the state-set of a. This construction is used to give a new characterization, in terms of structural closure properties, of classes of coalgebras that are defined by “observable” formulas, these being Boolean combinations of equations between terms that take observable values. It is shown that the E-construction can be replaced by a modification that restricts to ultrafilters whose members are definable in a. Both constructions are examined from the category-theoretic perspective, and shown to generate monads on the category of coalgebras concerned.
1. Introduction and Overview This paper continues a series [GolOlc,a,b]of articles on the equational logic and model theory of coalgebras for certain functors T : Set -+ Set on the category of sets. A T-coalgebra is a pair ( A ,a ) comprising a set A , thought of as a set of “states”, and a function a : A -+ T A called the transition structure. We study the case of functors T that are polynomial, i.e. constructed from constant-valued functors and the identity functor by forming products, coproducts, and exponential functors with constant exponent. Many data structures and systems of interest to computer science - such as lists, streams, trees, automata, and classes in object-oriented programming languages - can be modelled as coalgebras for polynomial functors [Rei95, Jac96, Rut95, RutOO]. This has motivated the development of a theory of “universal coalgebra” IRut95, RutOO], by analogy with, and categorically dual to, the study of abstract algebras. *Prepared using Paul Taylor’s diagrams package.
153
The article [GolOla] developed a type-theoretic calculus of terms for operations on polynomial coalgebras and explored its semantics. A special role is played by terms that take “observable” values. A polynomial coalgebra can be thought of as being constructed, using products and coproducts etc, from some fixed sets of observable elements given in advance. Computationally, the states of a coalgebra are regarded as not being directly accessible, but can only be investigated by performing certain “experiments” in the form of coalgebraic operations that yield observable values. Hence the emphasis on observable terms. It was shown in [GolOla]that Boolean combinations of equations between observable terms form a natural language of observable formulas for specifying coalgebraic properties. In particular such formulas give a logical characterization of the fundamental relation of bisimilarzty, or observational indistinguishability, between states of coalgebras: two states are bisimilar precisely when they satisfy the same observable formulas [GolOla, Theorem 7.21. The subsequent article [GolOlb]adapted the theory of ultrapowers to the context of polynomial coalgebras. Given a 3”-coalgebra a and an ultrafilter U , the standard theory of ultrapowers produces a structure au that is not a T-coalgebra. It was shown how to modify ’a to overcome this problem, by removing certain states. The result was the notion of the observational ultrapower a+ of a with respect to U.This notion was then used to give a structural characterization of classes of coalgebras definable by observable formulas: a class K is the class of all models of a set of such formulas iff K is closed under disjoint unions, images of bisimilarity relations, and observational ultrapowers [GolOlb, Theorem 7.11. The purpose of the present paper is to replace the a+ construction in this last result by a “Stone space like” construction that is intrinsic to a. We define the ultrafilter enlargement Ea of a as a new coalgebra whose states are certain ultrafilters on the state set of a. Ea is a homomorphic image of any observational ultrapower a+ that is sufficiently saturated, and we make use of this fact to transfer the analysis of a+ to Ea. In particular we use the version of LoS’s Theorem developed for a+ in [GolOlb]to study the conditions for truth in Ea at a state (ultrafilter) F . If pa is the set of states in a at which formula cp is true, then it transpires that
Ea,F
bcp
iff
(pa
EF,
which may be interpreted as saying that cp is true in Ea at state F iff it is true in a at a set of states that is “large in the sense of F” (see the Truth Lemma 3.3 below).
154
State-sets of the form cp“ may be called definable in a. The collection D e f , of all such definable sets is a Boolean algebra, and an alternative construction to Ea results by taking states to be ultrafilters of this algebra D e f , , rather than of the algebra of all subsets of a. The result is the definable enlargement Aa of a. Aa is a homomorphic image of Ea, and can be realized as the quotient of Ea by the bisimilarity relation. The category-theoretic nature of the Ea and A a constructions are investigated: in the final section of the paper we show that each gives rise to a monad structure on the category of coalgebras. A polynomial functor is monomial if it is constructed without the use of coproducts. A notion of ultrafilter enlargement for monomial coalgebras was developed in [GoldOlc]. Its theory is much simpler than the one described here: the presence of coproducts introduces considerable complexity associated with the partiality of certain “path functions” expressing the dynamics of the transition structure a. 2. Essential Background
A substantial conceptual framework and notational system is developed in [GolOla] and [GolOlb], all of which is essential to understanding the constructions and results given here. We now review this material, in order to make the present paper reasonably self-contained and accessible. 2.1. Polynomial Functors and Coalgebras
First, here is the notation for products, powers and coproducts of sets. For j = 1 and 2, rj : A1 x A2 -+ Aj is the projection function from the product set A1 x A2 onto Aj,i.e. rj (a1,a2) = aj. The pairing of two functions of the form f1 : A -+ B1 and f2 : A -+ B2 is the function (f1, f2) : A -+ B1 x B2 given by f ( a ) = (fi(a),f~(a)). The product of two functions of the form f1 : A1 -+ B1 and f2 : A2 -+ B2 is the function f1 x f2 : A1 x A2 -+ B1 x B 2 that maps (a1,a2)t o (fl(al),fi(a2)).Thus nj((f1 x f2)(2))= f j ( r j ( 2 ) ) . The coproduct A1 + A2 of sets Al,A2 is their disjoint union, with injective insertion functions ~j : Aj -+ A1 +A2 for j = 1 and 2. Each element of A1 A2 is equal t o ~j (a)for a unique j and a unique a E Aj. The coproduct of two functions of the form f1 : A1 -+ B1 and f2 : A2 -+ B2 is the function fl f2 : A1 A2 -+ B1 B2 that maps ~j(a)to ~j(fj(a)). The D - t h power of set A is the set AD of all functions from set D to A. The D-th power of a function f : A -+ B is the function f D : AD -+ BD having f D ( g ) = f o g for all g : D -+ A. The evaluation function
+
+
+
+
155
eval : AD x D + A has ewaZ(f,d) = f(d). For each d E D there is the evaluation-at-d function evd : AD + A having ev&) = eval(f, d ) = f(d). The symbol 0will be used for partial functions. Thus f : A 0B means that f is a function with codomain B and domain Dom f C A. Associated with each insertion ~j : Aj + A1 + A2 is its partial inverse, the extraction function ~j : A1 A2 -0 Aj having ~ j ( y = ) x iff ~ j ( x= ) y. Thus DomEj = ~ j A ji.e. , y E DomEj iff y = L ~ ( zfor ) some 2 E Aj. These extraction functions play a vital role in the analysis of coalgebras built out of coproducts. Observe that the coproduct f i f 2 of two functions has (fi + f 2 ) ( x )= ~ j ( f j ( ~ j ( 5for ) ) )some j . The identity function on a set A is denoted idA : A + A. If A is a subset of B , then A 9B is the inclusion function from A to B. Polynomial functors are formed from the following constructions of endofunctors T : Set + Set.
+
+
For a fixed set D # 0, the constant functor D has D(A) = D on sets A and D(f) = idD on functions f . The identity functor Id has IdA = A and Id f = f . The product TI x T2 of two functors has TI x T2(A)= T1A x T2A, and, for a function f : A + B , has TI x T 2 ( f )being the product function T l ( f )x T 2 ( f ): T1A x T2A + TiB x T2B. The coproduct Ti +T2 of two functors has TI +T2(A)= TlA+T2A, and for f : A + B , has TI T 2 ( f )being the coproduct function
+
+
T i ( f ) T 2 ( f ): T i A
+ T2A + TiB + T2B.
The D-th power functor T Dof a functor T has T D A= ( T A ) Dand , for f : A + B , has T D ( f )being the function ( T ( f ) ) D : ( T A ) D+ ( T B ) Dthat acts by g ++ T ( f ) o g .Thus T D ( f ) ( g ) ( d=) T ( f ) ( g ( d ) ) .
A functor T is polynomial if it is constructed from constant functors and Id by finitely many applications of products, coproducts and powers. Any functor formed as part of the construction of T is a component of T . A polynomial functor that does not have Id as a component must be constant. A T-coalgebra is a pair ( A , a ) comprising a set A and a function
A
a
T A . A is the set of states and a is the transition structure of the coalgebra. Note that A is determined as the domain Doma of a,so we can identify the coalgebra with its transition structure, i.e. a T-coalgebra is any function of the form a : Doma + T(Doma). A morphism from
156
T-coalgebra a to T-coalgebra ,6 is a function f : Doma + Domp between their state sets which commutes with their transition structures in the sense that p o f = T f o a , i.e. the following diagram commutes:
Doma
f
Domp
f is an isomorphism if it has an inverse that is also a coalgebraic morphism (or equivalently, if it is bijective). If Dom a C Dom p, then a is a subcoalgebra of p iff the inclusion function Dom a L) Dom /?is a morphism from a to p. More generally, if X is a subset of B, then there exists at most one transition p' : X + TX for which the inclusion X L) B is a 5"-morphism from p' to /3 [RutOO, Proposition 6.11. Thus it makes sense to talk about the set X being a subcoalgebra of p. For any morphism f : ( A ,a ) -+ ( B ,p), the image f ( A ) is a subcoalgebra of p, and if f is injective then coalgebra a is isomorphic to this image coalgebra [RutOO, Theorem 6.31. Every set {ai : i E I } of 2'-coalgebras has a disjoint union X I ai, which is a T-coalgebra whose domain is the disjoint union of the Dom ai's and whose transition structure acts as aj on the summand LjDomaj of Dom Crai. More precisely, this transition is given by ~ j ( a )I+ T ( ~ j ) ( a j ( a ) )with , the insertion ~j : Domaj + Dom CIai being an injective morphism making aj isomorphic to a subcoalgebra of the disjoint union (see [RutOO, Section 41).
2 . 2 . Paths and Bisinzulations
If ( A , a ) and (B,P) are T-coalgebras, then a relation R 5 A x B is a Tbisimulation from a to p if there exists a transition structure p : R -+ T R on R such that the projections from R to A and B are coalgebraic morphisms
157
from p to a and
p, i.e. the following diagram commutes:
TA
- TTl
TR
TTZ
TB
We may say that coalgebra p is the image of the bisimulation, or is the image of a under the bisimulation, if R is surjective, i.e. every member of B is in the image of R. Dually, a! is the domain of the bisimulation if R is a total relation, i.e. Dom R = A. A function f : A + B is a morphism from a to p iff its graph { ( a ,f ( a ) ) : a E A } is a bisimulation from a to p [RuttOO, Theorem 2.51: a morphism is essentially a functional bisimulation. When Doma C Domp, a is a subcoalgebra of /3 iff the identity relation on Dom a is a bisimulation from a to p. The above categorical definition of bisimulation appeared in [AM89]. It has a characterization [Her93, HJ981 in terms of “liftings” of relations R 2 A x B to relations RT C T A XT B . This in turn was transformed in [GolOla] to another characterization of bisimulations that uses the idea of “paths” between functors, an idea introduced in [JacOO, Section 61. Informally, the construction of polynomial T can be parsed into a tree of component functors. The paths we use are just the paths through this tree. Formally, a path is a finite list of symbols of the kinds ~ j ~ , j evd. , Write
P
p . q for the concatenation of lists p and q. The notation T S means that p is a path from functor T to functor S, and is defined by the following
conditions
T
-+ 0
--
T , where () is the empty path.
TI x TZ
nj .P Ej
.P
S whenever Tj
P
S, for j = 1,2.
P
TI TZ S whenever Tj S, for j = 1,2. evd.P P TD S for all d E D whenever T S.
It is evident that for any path T-S,
-
-
S is one of the components of T . P 9 Paths can be composed by concatenating lists: if TI TZand TZ T3, P.9 then TI T3.
-
1%
P
S induces a partial function PA : T A -0 set A, defined by induction on the length of p as follows. A path T 0
( ) A : TA
TA is the identity function idTA, so is totally de-
0
fined. 0
S A for each
(rj.p)= ~ p ~ o ~thecompositionofTldxT2A j , Thus x E Dom (7rj.p)~iff ~ j ( x E ) DOmpA.
rj
TjA
PA 0-
SA.
€, PA ( ~ j . p= ) ~p ~ o ~the j ,composition of TlA+T2A OL TjA 0SA. iff z E DomEj and &j(x)E DOmpA. Thus x E Dom (E~.P)A evd PA (evd.p)A = PA o evd, the composition of (TA)D T A 0SA. Thus f E Dom (evd.p)A iff f ( d ) E DOmpA.
-
Concatenation of paths corresponds to composition of functions, in the sense that (p.q)A = q A o P A . A path T-S is a state path if S = Id, an observation path if S = D for some set D, and a basic path if it is either. A T-bisimulation can be characterized as a relation that is “preserved” by the partial functions induced by state and observation paths from T . To explain this we adopt the convention that whenever we write “f(z) Qg(y)” for some relation Q and some partial functions f and g we mean that f ( z ) is defined iff g(y) is defined, and (f(a), g ( y ) ) E Q when they are both defined.
-
Theorem 2.1. [GolOla, Theorem 5.71 If A a TA and B P TB are coalgebras f o r a polynomial functor T , t h e n a relation R A x B is a T-bisimulation if, and only if, xRy implies
-
(1) f o r all state paths T
-P
Id, P A ( ~ ( X ) )R p ~ ( p ( y ) ) and ;
(2) f o r all observation paths T
P
0, p ~ ( a ( ~=) p) ~ ( P ( y ) ) .
0
The inverse of a bisimulation is a bisimulation, and the union of any collection of bisimulations from a to ,d is a bisimulation [RuttOO, Section 51. Hence there is a largest bisimulation from a to P, which is called bisimilarity. We denote this by N . States x and y are bisimilar, x y, when xRy for some bisimulation R between a and P. This is intended to capture the notion that x and y are observationally indistinguishable. N
2.3. Types, Terms, and Formulas
Fix a set 0 of symbols called observable types, and a collection { [ 01 : o E 0} of non-empty sets indexed by 0. [ 01 is the denotation of 0,and its members
1 59
are called observable elements, or constants of type 0. The set of types over 0, or 0-types, is the smallest set T such that 0 T,St E T and (1) if u1,m E T then 01 x 02, u1 +u2 E T; (2) if c E T and o E 0, then 0 + u E T.
A subtype of an Otype 7 is any type that occurs in the formation of 7. S t is a type symbol that will denote the state set of a given coalgebra. The symbol “0” will always be reserved for members of 0.o + u is a power type: such types will always have an observable exponent. A type is rigid if it does not have St as a subtype. The set of rigid types is thus the smallest set that includes 0 and satisfies (1) and (2). Each (Ortype (T determines a polynomial functor 101 : Set -+ Set. For o E 0,101 is the constant functor D where D = 601; IStJ is the identity functor Id; and inductively 101 X g 2 1
= ID11
X 1021,
= 1011
In1
For denotations of types, we write [.]A = I[d,[ s t ] A = A7 [‘TI
10
*
= lol’on.
for the set IalA. Thus we have
x u2]A = [01]A x [u2]A
[01+021A
[o
[.]A
1021,
(T]A
=[cl]A+iCJ2]A
= [.]A‘”’.
If u is a rigid type then 101 is a constant functor, so [.]A is a fixed set whose definition does not depend on A and may be written [ u ] . A r coalgebra is a coalgebra ( A ,a ) for the functor (71, i.e. a is a function of the form A -+ 1[T]A. To define t e r m s we fix a denumerable set V a r of variables and define a context to be a finite (possible empty) list
r = (vl : ul,. . . ,v, : a,> of assignments of Otypes ui to variables vi,with the proviso that v1, . . . , II, are all distinct. I? is a rigid context if all of the ui’s are rigid types. Concatenation of lists r and r’ with disjoint sets of variables is written I’,“’. A term-in-context is an expression of the form rDM:u,
which signifies that M is a “raw” term of type u in context I?. This may be abbreviated to r D M if the type of the term is understood. If u E 0, then the term is observable.
160
Figure 1 gives axioms that legislate certain base terms into existence] and rules for generating new terms from given ones. Axiom (Con) states
Coproduct Types
Figure 1. Axioms and Rules for Generating Terms
that an observable element is a constant term of its type, while the raw term s in axiom (St) is a special parameter which will be interpreted as the “current” state in a coalgebra. The rules for products, coproducts and powers are the standard ones for introduction and transformation of terms of those types. The raw term in the consequent of rule (Case) is sometimes abbreviated to case(N, M I ,Mz).
161
Bindings of variables in raw terms occur in lambda-abstractions and case terms: the v in the consequent of rule (Abs) and the wj's in the consequent of (Case) are bound in those terms. It is readily shown that in any term D M, all free variables of M appear in the list r. A ground term is one of the form 0 D M : u,which may be abbreviated to M : (T,or just to the raw term M. Thus a ground term has no free variables. Note that a ground term may contain the state parameter s, which behaves nonetheless like a variable in that it can denote any member of Doma, as will be seen in the semantics presented in Section 2.4. There exist ground terms of every type, as may be seen by induction on type formation. A term is defined to be rigid if its context is rigid. This entails that any free variable of the term is assigned a rigid type by I?, so its type is formed without use of St. Of course all ground terms are rigid. r-Terms
For a given Otype r , a r - t e r m is any term that can be generated by the axioms and rules of Figure 1 together with the additional rule
(r-n-1
I'DM:St rDtr(M) : r .
Note that from this rule and the axiom (St) we can derive the ground r-term
0 D tr(s) : r.
-
The symbol tr will denote the transition structure of a r-coalgebra
A
a
[ ? - ] A . If interpreted as a(.).
M is interpreted as the state
z of a , then tr(M) is
r-Formulas
An equation-in-context has the form I? D M1 x Mz where D M1 and I' D M2 are terms of the same type. A formula-in-context has the form I' D cp, with the expression cp being constructed from equations MI M Mz by propositional connectives. Formation rules for formulas are given in Figure 2, using the connectives iand A. The other standard connectives V, +, and H can be introduced as definitional abbreviations in the usual way. A formula 0Dcp with empty context is ground, and may be abbreviated to cp. A rigid formula is one whose context is rigid.
162
I Equations [
Weakening
(Weak)
I
r,
r,r' D 'p : at, r' D cp
where v does not occur in
r or r'
Connectives
Figure 2. Formation Rules for Formulas
A 7-formula is one that is generated by using only r-terms as premisses in the rule (Eq). An observable formula is one that uses only terms of observable type in forming its constituent equations. 2.4. Semantics of Terms and Formulas
A 7-coalgebra Q : A + ) r ) Ainterprets types a and contexts I' = ol,. . . ,vn : an) by putting [ a ] , = laJ(Doma)= [ [ T ] A , and
(vi :
11r1,= [ ~ i ] ,x ... x Iann,
[a],
(so is the empty product 1). The denotation of each 7-term r D M : a, relative to the coalgebra a , is a function
[r D M
:a]a : A x
[ria+
defined by induction on the formation of terms. For empty contexts,
Ax
[[a],
= A x 1 S A,
so we replace A x [ 0 ]la by A itself and interpret a ground term 0 D M : a as a function A + [ a ] , . The definition of denotations for terms given by the axioms of Figure 1, and the rule (-r-Tr),are as follows. Var : [ v : a D v : c 1, : A x [ u 1, + [ (T 1, is the right projection function.
Con:
[ 0 o c : o ], : A --t I[ o 1 is the constant function with value c.
163
St:
[8 D s : S t 1, : A -+ [ S t 1, is the identity function A -+ A. 7-TC I[ J?
D
t r ( M ) : T 1, : A x I[ r 1,
+ I[ r 1,
is the composition of the functions
The denotations of terms generated inductively by the rules of Figure 1are given by definitions that are standard in categorical logic (see [GoldOla, Section 41 or [Pitoo]).
Substitution of Terms The term N [ M / v ]is the result of substituting the raw term M for free occurrences of the variable u in N . The following rule is derivable: (Subst)
rDM:a
r,v:uDN:a'
D N [ M / v ]: u'
The semantics of terms obeys the basic principle that substitution is interpreted as composition of denotations [PitOO, 2.21. Because of the special role of the state set A, this takes the form
so that the following diagram commutes:
Substitution for the State Parameter The term " M I S ] is the result of substituting M for the state parameter s in N , according to the derivable rule (s-Subst)
rDM
I? D N : u' : St D N [ M / s ]: u'
164
For ground terms (I? = S), this takes the simple form
[wew i e .
I I N ~ ~ I= II,
0
Semantics of Formulas
A r-equation D M I M M2 is said to be valid in coalgebra a if the adenotations [ I' D M I le and [ r D M2 ]Ie of the terms r D Mj are identical. More generally we introduce a satisfaction relation a,x,yb r D cp, for r-formulas in r-coalgebras, which expresses that I' D cp is satisfied, or true, in a at state x under the value-assigment y E [I'lla to the variables of context r. This is defined inductively by a , Z , y k r D M l %y2 iff [ r t > M l ] I e ( ~ , y ) = [ r ~ M 2 ] I e ( ~ , y ) , a,Z,-YkrDDcp
iff not a , x , y krDcp,
a,x,ybrDcpiA(pz
iff a , x , y k r D D i a n d % Z , y k r D V 2 .
r D cp is true at Z, written a , z + r D cp, if a , x , y t= r D cp for all y E [r],. a is a model of r D cp, written a D cp, if a , Z, b r D cp for all states x E Dom a. In that case we also say that
D cp is
valid in the coalgebra a.
The following result is proven in [GolOla, Section 51.
Theorem 2.2. The class { a : a b D cp} of all models of a n observable formula is closed under domains and images of bisimulations, including domains and images of morphisms as well as subcoalgebras. If r D cp is rigid and observable, then its class of models is also closed under disjoint 0 unions. By substituting a ground term M for the state parameter s in given formulas we can produce formulas cp[M/s]that express the modal assertion that cp
165
will be true after execution of the state transition x +-+ I[M],(x) defined by M . This is the content of the following result.
Theorem 2.3. [GolOla, Theorem 6.51 If M is any ground term of type St, and r-coalgebra ( A ,a ) ,
'p
any ground formula, then in any
2.5. The Role of Observable Formulas Observable terms and formulas (and especially ground ones) play a role in the theory of polynomial coalgebras comparable to that played by standard terms and equations in the theory of abstract algebras. We record here some results that will be needed from [GolOla], concerning ways in which observable terms and formulas characterize structural aspects of coalgebras.
Theorem 2.4. [GolOla, Corollary 5.31 Let F D ' ~be any rigid observable r-formula. If R is a 17-1-bisimulationfrom a to P and xRy, then
In particular, i f f : A
+B
is a morphism from a to p, then for any x E A,
Consequently, iff is a surjective morphism,
Theorem 2.5. [GolOla, Theorem 6.71 A function f : A + B between r-coalgebras (A,a) and (B,P) is a morphism if, and only if, for all x E A ,
171-
(1) [ M ] , ( x ) = [ M ] p ( f( x ) )for all ground r-terms M of observable type; and (2) f ( [ M ] , ( x ) )= [ M ] I p ( f ( x )for ) all ground r-terms M of type St. 0
166
2.6. Defining Path Action and Bisimilarity The action of a path function is definable by a (ground) term, in the following sense.
-
Lemma 2.6. (Path Lemma) [GolOla, Theorem 6.11
For any path form
P
171
101
and variable v there exists a tr-free T-term of the v : ~ D p : a
such that f o r any 7-coalgebra ( A ,a ) and any x E A , z f a ( x ) E DOmpA then PA(+))
= I [ P W / ~ I iff( x ) .
Note that by the substitution rule (Subst), p [ t r ( s ) / v ]is a ground term of type g, since tr(s) is a ground term of type T . The term function [ p [ t r ( s ) / v ]l a has domain A , and so may not be identical to P A o a if P A is partial. This is only an issue when the path p includes an extraction symbol ~j (for otherwise p~ is total), but use of case allows the construction of observable terms that “discriminate” between the two summands of a coproduct [ T I ] A [r2] A and determine whether ~ A ( Q ( x ) is ) defined [GolOla, Section 61. For this to work it is necessary to assume that there is available at least one observable type p that is non-trivial in the sense that [ p ]I has at least two distinct members. This is a plausible assumption in dealing with notions that are to be discriminated by observable behaviour. Define a relation ~~p between the state sets of two r-coalgebras by putting
+
x
y iff every ground observable term M has [ M ] l f f ( x= ) [M ]p(y).
If r has at least one non-trivial observable subtype, then -ffp is a bisimulation from a to ,Ll [GolOla, Lemma 7.11. Moreover it proves to be the largest such bisimulation, giving a logical definition of bisimilarity. The precise situation is as follows. Theorem 2.7. [GolOla, Theorem 7.21 Let ( A , a ) and (B,,Ll)be 7-coalgebras, where T has at least one non-trivial observable subtype. Then f o r any x E A and y E B , the following are equivalent:
-
(1) x and y are bisimilar: x y . l? D cp f o r all rigid observable formulas ( 2 ) a ,z D cp igp,y
+
r D cp.
167
(3) a , x b cp i f fP, y b cp for all ground observable formulas 'p. (4) a , x /=M x N implies p, y /= M x N for all ground observable terms M and N . (5) [MICY(% =)[ M ]a(y) for all ground observable terms M , i e . x y. 0
2.7. Observational Ultrapowers Here we review the ultrapower construction of Section 4 of [GolOlb]. Let U be an ultrafilter on a set I . For each set A , there is an equivalence relation =U on the I-th power A' of A defined by
f
=U
g iff {i E I : f (i) = g ( i ) } E U.
Each f E A' has the equivalence class fu = {g E A' : f =u g}. The quotient set
AU = { f u : f E A'} is called the ultrapower of A with respect to U . A notation that will be useful below is to write f E u X , for X 5 A, when {i E I : f(i) E X } E U . We may also safely write f U E u X in this case, since in general f E U X iff g EU X whenever f =u g. There is a natural injection eA : A H AU given by eA(u) = tiu, where a E A' is the constant function on I with value a. The distinction between a and 8 is sometimes blurred, allowing A to be identified with the subset e A ( A ) of Au. A map 8 : A --f B has a U-lifting to Ou : AU -b BU where eU(f u , = (00 f ) u . This works also for a partial 8 : A 0B , providing a U-lifting Bu : AU 0Bu in the same way, with the proviso that f E Dom Bu precisely when f E U Dome, i.e. when {i E I : f ( i ) E Dome} E U . Moreover, U lifting commutes with functional composition: given also 7 : B 0C we have (7 o e)u = q' o Bu : Au 0Cu. be a T-coalgebra. The transition structure 01 Now let 01 : A + lifts to a function au : Au + I T ] ; , but this a' is not a T-coalgebra on Au since its codomain is [TI," = ( I T I ( A ) )rather ~ than [ T ] A V = I T [ ( A ~ ) . To overcome this obstruction it is necessary to remove some points from A U . The key to understanding which ones are to be retained is provided by considering the U-lifting of the a-denotation of a ground observable term M : 0. This is the function [ M ] : : Au + [onu. To act as a denotation for M it should assign values in 101, viewed as a subset of In other
168
words we should have
[MI:(%)
E e [ o ] = {8' : c E [ o ] }
[.Iu.
We are thus led to define an element z of AU to be obseruable if [ M ]:(%) E e [ o ] for every ground observable .r-term M : 0. If z = f U ,this means that for each such M there exists an observable element C M E [OD such that [ M ] : ( z ) = C M and ~ so
{i E I : [ M I m ( f ( i )= ) cM} E
Put A+ = {z E Au M : 0,
:
u.
z is observable}. For each a E A and any ground
so eA(a) is observable. Thus eA embeds A into A+, allowing us to view A+ as an extension of A. The definition of a IT)-transition structure a+ on A+ depends on the nature of the functor 1 ~ 1 ,which can be analysed in terms of paths 1r1-l~l. The definition of a+ is founded on the following technical result, whose proof proceeds by induction on the length of u.
-
Theorem 2.8. [GoOlb, Theorem 4.11
For any path 171 (PA o a)+ : A+ 0 diagram
P
1u1 beginning at
171
there exists a partial function
I[ CIA+with domain A+ nDom (PA o a)u such that the A* 0
PA
eA
-c
A+ 0
0
commutes wherever defined: i f a E Dom ( p ~ o athen ) eA(a) E Dom ( ~ A o ( Y ) + and
I+A(@A
o a ) ( a ) ) = (PA o a ) + ( e ~ ( a ) ) .
0
Now when u = r and p is the empty path, so that P A = idA, Theorem 2.8 gives a function a+ : A+ 0[ T ] A + whose domain is A+nDomau = A+, hence a+ is total, such that the following diagram commutes.
169
a+ is thus a r-coalgebra, which we call the observational ultrapower of Q over U . A feature of this construction is that the class of models of an observable formula is closed under observational ultrapowers. More strongly, if J? D cp is observable, then (Y
/= r D cp
if, and only if, a+
r D cp
[GolOlb, Corollary 5.31. This follows from a coalgebraic version of Log’s Theorem, which for our present purposes takes the following form.
Theorem 2.9. [GolOlb, Theorem 5.21 I f I’ D cp is an observable r-formula, and f u E A+, then a + , f U b r D c p if, a n d o n l y i f , { i E I : c u , f ( i ) ~ r D c p } E U .
0
Ultrapowers are used in first-order model theory to build extensions of structures that are “saturated” (full of elements). We make use of a kind of saturation notion, which is expressed by saying that Q+ is enlarging if the following property holds:
any collection S of subsets of A with the finite intersection property has a “nonstandard element in its intersection”. This element is a n x E A* such that f o r each X E S,x Eu X . Enlarging observational ultrapowers can be obtained by choosing a suitable ultrafilter U [GoOlb, Section 61. A structural characterization of logically definable classes of polynomial coalgebras can now be stated:
Theorem 2.10. [GoOlb, Theorem 7.11 If r has at least one non-trivial observable subtype, then for any class K of r-coalgebras, the following are equivalent. (1) K is the class of all models of some set of ground observable f o m u l a s . (2) K is the class of all models of some set of rigid observable formulas.
170
(3) K is closed under disjoint unions, images of bisimulations, and observational ultrapowers. (4) K is closed under disjoint unions, images of bisimilarity relations, and enlarging observational ultrapowers. 3. Ultrafilter Enlargements One of the main purposes of the present paper is to show that observational ultrapowers can be replaced in Theorem 2.10 by a "Stone space like" construction of the ultrafilter enlargement Ea, an object that is intrinsically determined by the coalgebra a itself. Assume from now that T is a type that has at least one non-trivial observable subtype. In any T-coalgebra A defines in A the "truth set"
(rDcp)"
a
= {z E A : a , x
[ T ] ] A , each
formula
D cp
r Dcp}
of all states at which the formula is true. Notice that for any morphism f : ( A ,a ) -+ ( B ,p), Theorem 2.4 states that if rDcp is rigid and observable, then in general z E (r D 9)" iff f(x) E (r D ( P ) ~and , so (r D 9)" = f-l
(rD cp)o.
An ultrafilter F on A will be called observationally a-rich, or more briefly just rich when a is understood, if it satisfies the following condition:
for any ground observable term M : o there exists some observable element C M E [.]I such that the truth set
(M M
cM)O
= {z E A :
pqa(x) =cM}
belongs to F . The element C M corresponding to M in this condition is unique, for if ( M M c)" and ( M M d)" belong to F then their intersection does too, hence is non-empty. But if z E ( M M c)" n ( M fi: d)", then c = [ M ] , ( z ) = d. The set of a-rich ultrafilters on A will be denoted EA. Each subset X of A determines the subset X E A of E A defined by
X E A= { F E E A :X E F } , and the map X I+ X E A preserves the Boolean set operations n, u and -. Members of E A can be constructed from the states of any observational ultrapower a+ : A+ + [ T ] A + of a over an ultrafilter U on some set I . For x E A+ define
@,v(z)= { X
A : z E,V X } .
171
Note that if x = f U , then for any X
X
E
A,
@ u ( f u ) iff {i E I : f(i) E X} E U.
It is standard theory that @ u ( f u )is an ultrafilter. But since f U is an observable element of A U , for each ground term M : o there is some C M E [on such that the set
{ i E I : [ M n a < f ( i )=) c M } = {i : f ( i ) E ( M M C
M ) ~ }
belongs to U , which implies ( M M C M ) ~E @ u ( f U )Thus . @ u ( f Uis) observationally a-rich, and we have a function @u: A+ -+E A .
Lemma 3.1. If a+ is enlarging, then @u: A+
-+ E A
is surjective.
Proof. Any F E E A has the finite intersection property, so by the enlarging property there is some x E A' with x EU X for all X E F . Since F is rich, for each ground M : o there is some C M E ion with x E U ( M M C M ) ~ This shows that x is observable, i.e. x E A + . But F C @u(x),so the maximality of F as a filter ensures that F = @u(x). 0 The definition of a .r-coalgebra structure E a on E A is a matter of similar complexity to the definition of a+ and requires an induction on paths from 17) to its component functors, similar to the proof of Theorem 2.8. We formulate the construction in a way that will enable us to transfer structure from any observational ultrapower of a to Ea by the maps @u. ~~
P
Theorem 3.2. For any path IT( (c(beginning at (TI there exists a partial function E ( P Ao a ) : E A o--c [ u ]IEA with domain equal to (DOmpA o a ) E A= { F E E A : Dom ( P A 0 a ) E F ) such that for any observational ultrapower A+ of A over an ultrafilter U the following diagram commutes wherever defined
.
172
The lengthy proof of this theorem is deferred to the next section. We proceed here to explore its consequences. In particular, when IS = r and p is the empty path with PA = idA, we get a commuting diagram
with the domain of E a being (Doma)EA. Thus D o m E a = E A , since Doma = A E F for all F E EA, so Ea is a total function. This gives the definition of E a as a r-coalgebra and the diagram shows that 9~is a morphism from a+ to E a . Ea is called the ultrafilter en2argement of the coalgebra a. If the diagram of Theorem 3.2 is composed with the square
of Theorem 2.8, the result is a commuting square
where
VA
= 9u
0
eA is the injection a
I+
{X
A : a E X}. In the case
173
that p is the empty path, this becomes
which shows that V A is an injective morphism a + Ea: of r-coalgebras that makes a isomorphic to a subcoalgebra of Ea. One of the benefits of Theorem 3.2 is that it enables the deep analysis of Log's Theorem (2.9) to be transferred to give information about truth conditions in the coalgebra Ea: Lemma 3.3. (Truth Lemma)
For any rigid observable formula I? D cp, and any state F E EA,
Proof. Let a:+ be an enlarging observational ultrapower of a with the associated map @U : a+ + E A being a surjection (Lemma 3.1). Given F E EA, choose fU E A+ such that F = @u(fU). Then by the invariance under a morphism of truth of a rigid observable
formula at a state (Theorem 2.4), we have that
Ea:,@u(fU)k
D cp
iff a+,fU
k
D cp.
By Log's Theorem 2.9, the latter condition is equivalent to
+r
{i : a , f ( i )
D cp)
E U,
i.e. to f' Eu (I' D cp)", and hence is equivalent to to by definition of @u.
(r D cp)"
E
@.v(fu) 0
We can now show that the class of all models of a rigid observable formula is closed under ultrafilter enlargements, and indeed is invariant under this construction:
Corollary 3.4. For any rigid observable I7 D cp,
174
+
Proof. If a I' D cp, then for each state F E EA, (r D cp)* = A E F , so Ea, F I? D cp by Lemma 3.3. This shows that Ea I? D cp. Conversely, if Ea r D cp, then a k I' D cp follows by Theorem 2.2, as a is isomorphic to a subalgebra of Ea by the morphism 734. Alternatively, a more streamlined proof is that for any enlarging a+, I' D cp by a Corollary to LOB'S Theorem mentioned a k I' D cp iff a+ earlier [GolOlb, Corollary 5.31, while a+ I ' D cp iff Ea k I? D cp by the last part of Theorem 2.4, which yields that validity in a coalgebra is invariant under the surjective morphism @u: a+ -+ Ea.
+
+
Now if M is a ground term of type St, then for any F E EA, the denotation value [ M ] E , ( F ) is a state of Ea, i.e. an observationally rich ultrafilter on A. The following is a useful characterization of members of this ultrafilter in terms of F .
Lemma 3.5. (State-Term Lemma) Let M be ground term of type St. Then for any F E EA, and any X
X E [MIE,(F)
A,
i f f [ M ] I i l ( X ) E F.
Proof. Given F E EA, let F = @u(fu),where @U : A+ -+ E A is the surjection given by an enlarging ultrapower of a. Now it was shown in [GolOIb, Theorem 5.1(4)] that for ground M : St, the denotation [MI,+ of M in a+ is just the restriction to A+ of the U-lifting [ M I : of the denotation [MI,: A -+ A. Thus [M]I,+(fU) = [ M ],"(.f"). Hence as the morphism @upreserves denotation values (The) ~ u ( [ M i , + ( f u ) )= ~ u ( ~ ~ ~ ~ ( f orem 2 . 5 ~ [)M~n E o l ( Q u ( f u ) = Then
x E wiE,(F) iffx E
@U([MI,U(fU))
from above
cU x
definition of
iff [ M n w )
@pu
fY Eu x
definition of [ M
iff {i E I : [ M ] , ( f ( i ) ) E X } E U
definition of E U
iff {i E I : f(i) E [ M
definition of [ M
iff ([MI,
O
(X)} E U
iff [ M ] I i l ( X )E @ u ( f U=) F
Corollary 3.6. For any X
1: 1;'
definition of @u.
A,
IIM],k(XEA) = ([M],l(X))EA.
0
'I75
Proof. For any F E EA, from Lemma 3.5 and the definition of X E A we get that
4. The Proof of Theorem 3.2
This section could be skipped over on a first reading of the paper, if the reader wishes t o continue at this point with the conceptual development. The proof of Theorem 3.2 proceeds by induction on the formation of the end-type 0 of the path p . In each case we define Dom E(PAo a ) to be the set (DompA o a ) E A= { F E E A : Dom ( P A o a ) E F } . But then if z E Dom ( P A W ) + , Theorem 2.8 implies that z E Dom ( p ~ o a ) ~ , which means that z E U Dom ( P A o a ) ,and therefore Dom ( P A o a ) E +U(Z), i.e. @u(z)E (DompA o a ) E Aas required. Thus the burden of the proof in each case is to define E ( P Ao a ) itself in such a way that E(PA0 a ) ( @ u ( z = ) ) ICT/@U((PA o a)+(x)):
A+ 0
*EA 0
TOunderstand the proof it is necessary to know the definition of the function (PA o a)+. This will be stated in each case, but it might be beneficial if the reader had access to the proof of Theorem 2.8 given in [GolOlb, Theorem 4.11. The induction begins with the base cases of observable types and the type St.
I Case
CT
1
E 0 Let D = [u].Then the above diagram in this case is
(PA
i
a)+
176
Let Mp = p[tr(s)/w]be the ground term of type r~ given by the Path Lemma 2.6. For F E Dom E ( ~ oAa ) , let E ( ~ oAa ) ( F )be the unique d E D such that the truth set ( M , M d)" belongs to the rich ultrafilter F . Now if z = fU E Dom ( P A o a)+,then as 2 is observable, there is some c E D such that [ M p ] z ( z = ) EU. In the proof of Theorem 2.8, ( ~ A o Q ) + ( x ) is defined to be this c. But now
{i E I : f(i) E ( M p M c)"} = {i E I : [ M P ] " ( f ( i )= ) c} E
u,
so z E U ( M p x c)". Hence ( M , M c ) E~ @ ~ ( zimplying ), that c i s the value of E ( ~ oAa ) at @u(z),i.e. ( P A o a)+(.) = E ( ~ oAa ) ( @ u ( 2 )as ) , required for the diagram to commute. Here
1r~1is
the identity functor Id, so the diagram becomes
A+ >
Define a unary operation putting
[pA] :
PA
+ P A on the powerset
P A of A by
[pA]X = { a E A : a E Dom ( p o ~a ) implies p ~ ( a ( a )E)X } = -Dom ( P A 0 a ) U { u E Dom ( P A 0 a ) : p ~ ( a ( a )E) X } .
It is straightforward to verify that [pA](Xn y ) = [pA]Xn [ p A ] y [pA](x u y ) = [pA]Xu [ p A ] y [pA]@= -Dom ( p 0~a). For any ultrafilter F on A, let Fp be the inverse image of F under
6) (ii) (iii) [pA]:
Fp = { X C A : b A ] X E F } . Since F is a filter, it follows from (i) that X f l Y E Fp iff X , Y E Fp, which means that Fp is a filter. Then if F E (DOmpA ~ a )since ~ Dom ~ (, p oa) ~ E F we get [pA]@4 F by (iii), so 0 4 Fp, and therefore Fp is proper. But (ii) implies that Fp is prime ( X u Y E Fp only if X E Fp or Y E Fp),so altogether Fp is an ultrafilter on A in this case. Moreover, if F is rich, then so is Fp. For, given a ground observable term N : o, consider the term N [ M p / s ]: o, where M p = p[tr(s)/w]: S t as in
177
Case u = u1 x
02
In this first inductive case we make the hypothesis that
A+ 0
@lJ
EA 0
178
fulfils Theorem 3.2 for any ultrafilter U . NOW PjA = rj o p A , where xj projects [(TI ] A x I [ ( T ~ ] onto A [cj] A , so as rj and a are total, Dom (gA o a ) = Dom (PA o a). Thus if Dom ( p 0~a ) E F E EA, then by induction hypothesis E V A o a ) ( F )is defined for j = 1,2, so we can define
a ) ( F )= ( E ( p i a > ( F ) , E ( p i a >(F ))-
E(pA
This yields the diagram @U
A+ (PA
I[ O1 ] A +
I
EA
IEbA O a) 1.1
x
021@U
102 ]A+
1.1
]EA
x
[a2 ]EA
In this case of (T = (TI x ( ~ 2 (, p o ~a)+ is defined to be the pairing function ( ( p i o a)+,( p i o a)+),so if z E Dom ( p o ~a)+,then for j = 1 , 2 , rjIE(pA
a>(@U(z)>I
= EVA O .>(@U(X)>
definition of E ( ~ oAa )
= Igjl@U((P”A
second-to-last diagram
= l ( T j l @ U ( X j ( ( p A a)+(.)))
definition of
( P A o a)+
= xj[l(~1x CTZ~@U((PA 0 a)+(.))] definition of /elx ezl. Hence E ( p A 0 a ) ( @ U ( x ) ) = 101
02l@U(bA
a)+(.)),
making the last diagram commute as required.
I Case
(T
= 01
+
(TZ
1 Assume Theorem 3.2 holds for
P.Ei
01
and
02.
This time we
lujl and, by the induction hypothesis, have define pl to be the path 17) a partial function E($A o a ) such that the same diagram
A+
@U
* EA
179
fulfils Theorem 3.2. But now = ~j o p ~where , ~j is the (partial) extraction from IT^ ] A [ ~ ] 2A to [Uj] A , and Dom ( P A o a ) is the disjoint union of Dom ($\o a ) and Dom ( p i 0 a ) . Thus if Dom ( P A o a ) E F E EA, then Dom (gA0 a ) E F for exactly one j, and we define
+
for this j , where ~j is the insertion of yields the diagram
[uj ]EA
into
[a1] E A
@U
A+ (PA
a)(F))
a ) ( F )= L j ( E ( &
E(PA
+
EA ]E @ A a )
a)']
[gl ]A+
+ [ U Z ] E A . This
+ [g2]IA+
Iff1
+ U21@U
[Ul]EA+[u2]EA
In this coproduct case, Dom ( P A o a ) + is the disjoint union of Dom ( p i o a ) + and Dom ( p i o a)+,and ( p o ~a)+(.) = ~j((g~ o a)+(.)) for the unique j such that (gAo a)+(.) is defined. Then
making the last diagram commute as required.
1-
-
Assume the Theorem holds for u. Let D =
from the path
"2 IuI
(TI
P
=. 01 we obtain, for each
(0
171 and, by hypothesis on the diagram
A+
U ,a
d E
Then
D ,the path p d
=
partial function E ( p i o a ) such that
@U c
EA IEbi
(Pi [CIA+
1.1.
l4@u
I["]IEA
oa)
180
fulfils Theorem 3.2. Here p$ = e v d O ~ A with , e v d : [On:: + [ o ] A , so as e w d and a are total, Dom ( p i o a ) = Dom ( P A o a ) . Thus if Dom ( p 0~ a) E F E EA, then by induction hypothesis on O , E(p2 o a ) ( F )is defined for all d E D ,so we can define E ( ~ oAa ) ( F )as a function of type D + [ O ~ E Aby putting
E(pA a ) ( F ) ( d )= E(pi a ) ( F ) . This yields the diagram @U
A+
* EA
definition of E ( ~ oAa ) second-to-last diagram definition of ( p o~a)+. definition of
E(pA
a ) ( @ U ( x )= ) 10
*
Ol@U((PA
10
01.
j
a)+(x)),
making the last diagram commute as required. This completes the inductive proof of Theorem 3.2.
0
5. Definable Enlargements We now consider a modification of the ultrafilter enlargement construction. This will produce a natural quotient of the coalgebra Ea by focusing on the truth sets (pa
= {X E A : a , x
9)
181
of ground observable formulas cp. Such sets may be called definable, and the collection
Def
" = {cp"
: cp
is ground and observable}
is a Boolean algebra of subsets of A. This follows because in general @ n cpz = A 9): and A - pa = (-wp)", so Def" is a subalgebra of the powerset Boolean algebra PA. Now let A A be the set of all observationally rich ultrafilters of the Boolean algebra Def O . Hence a member of A A is a collection of definable sets. Note that the sets ( M NN CM)" required for the definition of "observationally rich" are all defined by ground observable formulas, so such sets belong to Def ". Let 6, : E A + A A be the restriction map taking each F E E A to
(cpr
8,(F) = F n Def
= { p a : cpa E F } .
It is readily checked that B,(F) belongs to A A when F E EA. Moreover, 8, is surjective: for any H E AA, H has the finite intersection property so extends to an ultrafilter F on A which is rich because H is rich. Then F E E A and H g 8,(F), so H = 8,(F) as H is a maximal filter in Def ". Theorem 5.1. 6,(F) = 8,(G) if, and only if, F and G are bisimilar states in the coalgebra Ea. Proof. 8,(F) = B,(G) iff F and G contain the same sets of the form cpo with cp ground and observable. By the Truth Lemma 3.3 this means precisely that Ea, F I= cp iff E a , G cp for all such cp. But by Theorem 0 2.7(3), this holds iff F and G are bisimilar. Now let R = { (F,G) : F, G E E A and F and G are bisimilar}. Since R is a I+bisimulation (the largest one), there exists a transition p : R 3 [ T I E such that the diagram
R
commutes for j = 1 and j = 2.
182
Lemma 5.2. 8,(F)
where 1T18, : [ T ] E A to e, : E A + AA.
= 8,(G) implies ITlO,(Ea(F)) = ITlO,(Ea(G)), the result of applying the functor 17-1
+ [ T ] A A is
Proof. From the last diagram, for j = 1 , 2 ,
ITl8, o
so as
IT(
1~1rjo
p = 1T18,
0
Ea 0 rj,
o rj) o
p = 11 .8,
o
Ea o rj.
is a functor, ITl(8,
But Theorem 5.1 states that (F,G) E R iff 8,(F) = 8,(G), so the functions 8, o rland 8, o 7r2 from R to A A are identical, hence by the last displayed equation,
ITp,
E~ rl = ITp, E~
Tz.
Thus if 8,(F) = 8,(G), then (F,G) E R with
1T18, o Ea o r1 ( F ,G) = 1718, i.e. I+,(Ea(F))
0
Ea 0 r 2 ( F ,G) ,
= ITlO,(Ea(G)) as desired.
0
Theorem 5.3. There is a unique function Aa : A A 4 I[T]AA making the following diagram commute.
EA Ea
&Y
AA j Aa
Proof. Define Aa by putting Aa(e,(F)) = ITlO,(Ea(F)). Lemma 5.2 ensures that this is well-defined. Since 8, is surjective, the domain of Aa is AA. The definition of A a makes the diagram commute and is the only definition that can do so. 0 This result defines Aa as a .r-coalgebra and, importantly, makes 8, a surjective morphism from Ea to Aa. Aa is the definable enlargement of a. Theorem 5.1 states that the kernel of 8, is the bisimilarity relation on Ea, so Aa is isomorphic to the quotient of Ea by bisimilarity. Hence Aa is a simple coalgebra, i.e. itself has no proper quotients [RutOO, Proposition
183
8.21. In Aa itself, bisimilar states are equal. That also follows from Theorem 5.1, since bisimilarity is invariant under morphisms, so 8,(F) and B,(G) are bisimilar in A a precisely when F and G are bisimilar in E a , i.e. precisely when B,(F) = B,(G). The morphism 8, can be used to transfer the Truth Lemma 3.3 for E a , and its Corollary 3.4, to the corresponding results for Aa: Theorem 5.4. Let
D cp be a rigid observable formula.
+
(1) For any G E AA, Aa, G I' D cp i f f (2) a + r D c p iff A a k I ' D c p .
(rD cp)"
E G.
Proof.
(1) Given GI choose F E E A with G = B,(F). Then as 8, is a morphism, r D cp, which Theorem 2.4 yields that Aa, B,(F) b I? D cp iff Ea, F G by the Truth Lemma 3.3. in turn holds iff (I' D cp)* (2) From Corollary 3.4 we already know that a rDcp iff Ea rDcp. But as the morphism Oa is surjective, Theorem 2.4 yields that Ea b r D cp iffAa+I'rcp. 0 The morphism 6 , can also be used to transfer the State-Term Lemma 3.5 and its Corollary 3.6 to Aa: Lemma 5.5. Let M be ground term of type S t and let X E Def,. for any G E EA,
x E [M]A,(G)
Then
iff [M]G1(X) E G.
Consequently,
w]il,(xAA) = (IIMn;l(X))AA, where in general Y A A= {G E A A : Y E G}. Proof. Note first that if X = cp", then Theorem 2.3 states that I[ M ]la (z) E X iff 5 E cp[M/s],, so [ [ M ] ; l ( X )= cp[M/sIa,showing that [ M ] ; ' ( X ) is also definable. Now let G = B,(F) with F E EA. Then as 8, is a morphism, I[MDA,(G)
= e,(wiEa(F))=
iwEmn D ~ L ,
so as X is definable, X E [M]ld,(G) iff X E I [ M ] E , ( F ) , which holds iff [ M ] l ; l ( X ) E F by Lemma 3.5. But [ M ] ; ' ( X ) E F iff [ M ] ; ' ( X ) E G,
since [ M ] ; l ( X ) is definable as we just saw.
184
The rest of the Lemma then follows straightforwardly.
0
It follows from Theorem 5.4(2) that the class of all models of a rigid observable formula is closed under definable enlargements. In fact, in the structural characterization of such model classes set out in Theorem 2.10, observational ultrapowers can be replaced by ultrafilter enlargements, or by definable enlargements. To see this, first consider a class K of coalgebras that is closed under images of bisimulations. Then in particular it is closed under domains and images of coalgebraic morphisms, which means that for any surjective morphism f : a ++ p we have a E K iff /3 E K . This follows because the image of f is the image of the bisimulation Rf (the graph of f ) , while the domain of f is the image of the inverse relation RT', which is also a bisimulation. Now for any T-coalgebra a , if a+ is an enlarging observational ultrapower of a we have surjective morphisms
Thus if K is closed under images of bisimulations, and contains one of these three coalgebras, then it contains the other two as well. This observation, together with the equivalences of Theorem 2.10, yields the following extension of that Theorem.
Theorem 5.6. If T has at least one non-trivial observable subtype, then f o r any class K of T-coalgebras, the following are equivalent. (1) K is the class of all models of some set of rigid observable formulas. (2) K is closed under disjoint unions, images of bisimulations, and ultrafilter enlargements. (3) K is closed under disjoint unions, images of bisimulations, and definable enlargements. 0 6. Monads From Enlargements
In this section a category-theoretic perspective on coalgebraic enlargements is developed. The operation of assigning to each set A the collection of all ultrafilters on A gives rise to a categorical structure on the category Set of sets and functions that is known as a monad or triple (see [ML71, Chapter VI] or [Man76]). In a similar way, the Ea construction gives rise to a monad on the category r-Coalg of r-coalgebras and their morphisms.
185
For any morphism f : ( A ,a ) + ( B ,p) of r-coalgebras, define a function
Ef on EA by putting E f ( F ) = {Y
B : f-lY E F}.
Lemma 6.1. Ef is a morphism (EA,Ea) -+ ( E B ,ED). Proof. It is standard theory that E f ( F ) is an ultrafilter on B whenever F is an ultrafilter on A. To show it is observationally rich we use the fact, from the second sentence of Theorem 2.4, that for any ground observable formula cp, we have 2 E cp" iff f(z)E cpa in general, and so cp" = f-lcpo. For any ground observable term M : 0, a-richness of F implies that ( M M c)" E F for some c E [o]. Thus f-l(M M c)p E F by the last observation, and so ( M M c)O E Ef ( F ) . This shows that E f ( F ) is a ,&rich ultrafilter, so that Ef is indeed a function from E A to EB. Then to show Ef is a morphism it suffices, by Theorem 2.5, to show that for any F E E A and any ground term M , (1) [ M ] E , ( F ) = [ M ] E a ( E f ( F )if) M is observable; and (2) E ~ ( [ M I ~ , ( F )=) [ M n E , w ( F ) ) if M is of type
st.
+
For (I), let [ M ] E & ( F = ) c. Then Ea,F M M c, so ( M M c)" E F by the Truth Lemma 3.3. It follows as above that ( M M c)a E E f ( F ) ,hence E P , E f ( F ) M M c by 3.3 again. Thus
[ M ] E p ( E f ( F )= ) c = [M]IEa(F)For (2), as f is a morphism Theorem 2.5(2) states that the diagram
A-B UMI"1
f lfMla
A-B commutes. Hence for any Y C B ,
[ M];l(f-lY) Then
= f-l([M],lY).
186
y E Ef(IIM]Ea(F)) definition of E f
iff f - ' Y E [ M ] E , ( F ) iff I[ M
1;'
State-Term Lemma 3.5
( f -'Y) E F
iff f - l ( [ M ] p l Y )E F
from above
iff IIM]plY E E f ( F )
definition of E f
iff Y E [ M ] E o ( E f( F ) ) State-Term Lemma 3.5. Since this holds for all Y C B, (2) follows.
0
Ea and f I+ Ef provide It is now readily seen that the assignments a a functor E : T-Coalg + T-Coalg on the category of .r-coalgebras. Theorem 6.2. The morphisms V A : a + Ea are the components of a natural transformation 77 from the identity functor on 7-Coalg to the functor E. Proof. This amounts to the claim that for any morphism f the diagram
A
77A
EA
77B
lEf
B EB commutes in Set. But it is a simple set-theoretic calculation to show that E f ( q A ( 2 ) ) = V B ( f (2))for all 2 E A. Composing the functor E with itself gives the functor EE on .r-Coalg that assigns to each coalgebra ( A ,a ) a coalgebra ( E E A ,E E a ) whose states are the Ea-rich ultrafilters on EA. A function pa is defined on E E A by putting pa@) = { X
C A :X E AE P},
where X E A = { F E EA : X E F } , as in Section 3. Note that the notation pa is preferable to P A , since the definition depends on E A and hence on a. By contrast, the definition of 7 7 depends ~ only on the set A.
Theorem 6.3. The functions pa are the components of a natural transformation p from E E to E .
187
Proof. First it must be shown that pa is an arrow in T-Coalg (a morphism) from EEa to Ea. The fact that the map X e X E A preserves the Boolean set operations ensures that for each ultrafilter p E E E A , p Q ( p )is an ultrafilter on A. Moreover p a ( p ) is a-rich: for any ground term M : o there is some c E I[.] with ( M M c ) E p~, and ~ then by the Truth Lemma 3.3, { F E E A : ( M M C E) ~F } = { F : E a , F
M
M C }
= ( M e c ) ~ "E p ,
so ( M M c)" E p L , ( p )by definition of p a . This shows that p a ( p ) is rich, so pa is a function from E E A to EA. To show that pa is a morphism we apply Theorem 2.5, as in the proof of Lemma 6.1, this time showing that for any p E E A and any ground term M,
For (I), there exists an element c such that ( M NN c ) E ~p and ~ ( M "N c)" E pol@)as in the previous paragraph. By the Truth Lemma these imply that E E o , p M M c and Ea, p m ( p ) M M c, so that
For (Z), we reason that for any X C A ,
x E P=([M]IEEa(P)) iff X E A E [ M ] E E " ( ~ ) definition of pa iff I[ M ] I , k ( X E A ) E p
State-Term Lemma 3.5
iff ( [ M ] , ' ( X ) ) E AE p
Corollary 3.6
iff [ M ] ; ' X E p a ( p )
definition of p a
Lemma 3.5. iff X E [ M ] ~ ~ ( p ~ ( State-Term p ) ) Thus pa ([ M ] I E E " ( ~ ) ) = [ M ]Ea(pa( p ) ) ,completing the proof that pa is a morphism in .r-Coalg. Finally, to show p is natural it must be shown that the diagram
188
EEa
Pa
Ea
commutes in .r-Coalg whenever f is a morphism from a to P. This requires that
EEA
--kL.+ EA
-+
E E B PP EB commutes in Set, where A and B are the state sets of a and P. The proof of this is set-theoretic, requiring no further coalgebraic analysis, and is essentially part of the standard theory of ultrafilters [Man76, Section 1.31. The details are left to the interested reader, who would find it useful to first show that for any Y g B , +
( f - ' Y ) E A= ( E f ) - l ( Y E B ) .
0
The triple ( E , q , p )forms a monad on the category .r-Coalg. In addition to the naturality of 77 and p (Theorems 6.2 and 6.3), this means that for any .r-coalgebra ( A ,a ) the following diagrams commute.
-
EEEa EPa EEa
Ea
- EVA E E a
77EC-V
Ea
Demonstration of this reduces to showing commutativity of the corresponding diagrams in Set that result from replacing Ea by EA. Again these are standard ultrafilter calculations that need not be reproduced here. The reader who is interested to check the details would find it useful, in the case of the left diagram, to first show that for any X 2 A, &l(XEA) = (XEA)EEA.
189
The Definable Case The construction a c) A a also gives rise to a monad on .r-Coalg. First of all, A extends to a functor on .r-Coalg that assigns to each morphism f : ( A ,a ) -+ ( B ,p) the function Af : A A -+ A B having
A f ( G ) = {Y E D e f p : f - l Y E G } . The proof that O f is a morphism from A a to A p is similar to the proof that Ef is a morphism, using results 5.4 and 5.5 in place of 3.3 and 3.5. It is readily seen that the diagram
Ea
Qa
Aa
commutes, so the morphisms 0, are the components of a natural transformation 0 from E to A . A function q,d : A -+ A A is defined by q,d(z) = {X E D e f , : z E X } = e,(qA(Z)). Then q,d is a morphism from a to A a , being the composition of the morphisms q~ and 0,. The 7:’s are the components of a natural transformation from the identity functor on .r-Coalg to A , the composition of q and 8. Note that, unlike V A , q,d need not be injective: in general q,d(z)= q,d(y) iff z and y satisfy the same ground observable formulas in a , which holds iff z and y are bisimilar (Theorem 2.7). Thus q t is injective precisely when bisimilar states in a are equal. A natural transformation p A : A A -+ A is given by defining p,d(p) = { X E Def
:X A A E p } ,
where X A A = {G E A A : X E G } = Q , ( X E A ) (see Lemma 5.5). The proof that p t is a morphism A A a -+ A a is analogous to the proof that pa : EEa -+ Ea is a morphism. The triple ( A ,q A ,p A ) forms a monad on 7-Coalg, but one of a special kind, as the functor A is “idempotent up to isomorphism”, in the sense that A a and A d a are isomorphic. A “logical” explanation of this is that if cp and $ are ground observable formulas then by Theorem 5.4(2),
a+cp++$ iff AaI=cp*$,
190
and so 'pa = t+!Piff cpA" = Q"". Thus the map cp" H cpAa is a well-defined of definable subbijection between the Boolean algebras Def and Def sets of a and A a , respectively. Moreover this map is a Boolean isomorphism and gives a bijection between the sets of a-rich ultrafilters of Def L+ and taking G E A A to (9"": 'pa E G} E A A A . Aa-rich ultrafilters of Def This gives the isomorphism A a Z A A a . But the version of the Truth Lemma for A given in Theorem 5.4(1) shows that
cp"EG
iff GEcpAo,
so the isomorphism is the map G H (9"": G E cp""} = &(G). In other words, this isomorphism is just the component 77;"
: Aa
+Ada
of the natural transformation vA. Another proof that this component is a bijection follows from the observations firstly that is injective because A a is a simple coalgebra in which bisimilar states are equal, and secondly that is surjective because for anyp E A A A the set G = (9": cpA" E p } is a rich ultrafilter of Def o1 with q&(G) = p. It is noteworthy that part of this monad structure on A is the property that the diagram
772"
772,
A
Aa
A AAa
Aa commutes, so in fact the component p t of the natural transformation p A is itself the inverse of the isomorphism &, and hence is also an isomorphism. A monad on a category has an associated category of algebras. In the case of the A-monad, an algebra is a pair (a,f ) with f : A a + a a morphism for which the following commute: A
Ada
A
Aa
Af l -If
a
77"
Aa
f
Aa
a
a
191
But for an idempotent monad like A , in which the components p t are all isomorphisms, any such algebra ( a , f ) has f an isomorphism [Bor94, Proposition 4.2.31. For the ultrafilter monad on the category Set, the associated category of algebras is isomorphic to the category of compact Hausdorff topological spaces and continuous functions - this is Manes' Theorem, see [Man761 or [Joh82, Section I11 21. It would be of interest to know whether this situation lifts from Set to .r-Coalg, replacing the ultrafilter monad by the monad of E . Is there some topology that can be imposed on polynomial coalgebras that identifies a natural class of topological coalgebras isomorphic to the category of E-algebras f : Ea + Q! ? References Peter Aczel and Nax Mendler. A final coalgebra theorem. In D. H. Pitt et al., editors, Category Theory and Computer Science. Proceedings 1989, volume 389 of Lecture Notes i n Computer Science, pages 357-365. Springer-Verlag, 1989. Bor94. Francis Borceux. Handbook of Categorical Algebra 2. Categories and Structures. Cambridge University Press, 1994. GolOla. Robert Goldblatt. Equational logic of polynomial coalgebras. In Advances in Modal Logic, volume 4, World Scientific (to appear). Manuscript available at http: //wuw .mcs.v u w .ac .nz/'rob GolOlb. Robert Goldblatt. Observational ultrapowers of polynomial coalgebras. Manuscript available at http: //www .mcs.v u w .ac .nz/'rob GolOlc. Robert Goldblatt. What is the coalgebraic analogue of Birkhoff's variety theorem? Theoretical Computer Science, 266:853-886, 2001. Her93. Claudio Hermida. Fibrations, Logical Predicates and Indeterminates. PhD thesis, University of Edinburgh, 1993. Techn. rep. LFCS-93-277. Also available as Aarhus Univ. DAIMI Techn. rep. PB-462. Claudio Hermida and Bart Jacobs. Structural induction and coinducHJ98. tion in a fibrational setting. Information and Computation, 145:107-152, 1998. Jac96. Bart Jacobs. Objects and classes, coalgebraically. In B. F'reitag, C. B. Jones, C. Lengauer, and H.-J. Schek, editors, Object-Orientation with Parallelism and Persistence, pages 83-103. Kluwer Academic Publishers, 1996. JacOO. Bart Jacobs. Towards a duality result in coalgebraic modal logic. Electronic Notes in Theoretical Computer Science, 33, 2000. http: / / u u w . AM89.
elsevier.nl/locate/entcs.
Joh82. P. T. Johnstone. Stone Spaces. Cambridge University Press, 1982. Man76. Ernest G. Manes. Algebraic Theories. Springer-Verlag, 1976. ML71. Saunders Mac Lane. Categories for the Working Mathematician. Springer-Verlag, 1971.
192
PitOO.
Rei95.
Rut95.
RutOO.
Andrew M. Pitts. Categorical logic. In S. Abramsky, D. M. Gabbay, and T. S. E. Maibaum, editors, Handbook of Logic in Computer Science, Volume 5: Algebraic and Logical Structures, chapter 2. Oxford University Press, 2000. Horst Reichel. An approach to object semantics based on terminal co-algebras. Mathematical Structures in Computer Science, 5:129-152, 1995. J.J.M.M. Rutten. A calculus of transition systems (towards universal coalgebra). In Alban Ponse, Maarten de Rijke, and Yde Venema, editors, Modal Logic and Process Algebra, CSLI Lecture Notes No. 53, pages 231-256. CSLI Publications, Stanford, California, 1995. J.J.M.M. Rutten. Universal coalgebra: a theory of systems. Theoretical Computer Science, 249(1):3-80, 2000.
193
A LAYERED APPROACH TO EXTRACTING PROGRAMS FROM PROOFS WITH AN APPLICATION IN GRAPH THEORY
JOHN JEAVONS AND BOLIS BASIT Department of Mathematics and Statistics, Monash University, Australia E-mail: j [email protected], [email protected] AND
IMAN POERNOMO AND JOHN N.CROSSLEY School of Computer Science and Software Engineering, Monash University, Australia E-mail:{ihp,jnc}@csse.monash.edu.au
Abstract In this paper we describe our system for automatically extracting %orrect” programs from proofs using a development of the Curry-Howard process. Although program extraction has been developed by many authors (see, e.g., [7,3, lo]), our system has a number of novel features designed to make it very easy to use and as close as possible to ordinary mathematical terminology and practice. These features include (1) the use of Henkin’s technique from [8] to reduce higher-order logic to many-sorted (first-order) logic, (2) the free use of new rules for induction subject to certain conditions, (3) the extensive use of previously programmed (primitive) recursive functions, (4) the use of templates to make the reasoning much closer to normal mathematical proofs, and (5) an extension of the technique of the use of Harrop formulae to classically true formulae (cf. the footnote on p. 101 in Kreisel [ll]).
As an example of our system we give a constructive proof of the wellknown theorem that every graph of even parity, that is non-trivial in the
194
sense that it does not consist of isolated vertices, has a cycle. Given such a graph as input, the extracted program produces a cycle as promised. 1. Introduction
The well-known Curry-Howard isomorphism (see e.g. Howard’s original paper [9] or Crossley and Shepherdson’s paper [6] explicitly extending this to ordinary first order logic), produces a term of a lambda calculus from a (constructive) proof of a formula. This technique can be used to give a program that computes the constructive content of the formula. Thus, in arithmetic a constructive proof of a formula of the form Vx3yct.(x,y ) yields an algorithm for computing a function f such that a ( A , f ( A ) ) holds for every natural number n. ( A is the numeral for n.) In this paper we present an extension of the Curry-Howard isomorphism to a novel and expandable, first order, many-sorted, predicate calculus. Amongst other features, this logic also allows us to use previously programmed functions (and predicates), see below. The extension to a many-sorted calculus allows us to extract programs over different sorts. This has previously been done successfully in various higher order systems. Our approach avoids the use of higher order logic. It is well known that the programs extracted from full proofs in formal logic are immensely long both in size and in running time. We therefore introduce a number of novel features into our system. These are designed to mirror, as far as possible, normal mathematical practice. Besides the formal logical theory we also have a computational type theory. This computational theory is used to admit the use of pre-programmed functions and predicates. These functions (and predicates) can even be ones that we have just produced from our system. This is what we mean by “layering”. Moreover, we are able to retain a modularity between the computational type theory and the logical type theory of the Curry-Howard isomorphism. The interactions between the two are taken care of by our Curry-Howard protocol (see Section 3.1). These notions allow us to (1) (easily axiomatize and) use pre-programmed functions in our proofs in order to reduce the complexity and run-times of our programs, (2) retain a logic that is first order, (3) investigate and describe constructive proof “idioms” (analogous to programming “idioms” or “patterns”), and (4) define a protocol between programs and logic.
195
We have built a software system, written in ML and currently called proofEd, as an implementation of our system.a It has a UQXoutput feature, so that we can easily include proofs written in proof Ed in a document
such as the present paper. We demonstrate the system by using a constructive proof that every even parity graph contains a cycle and extracting a program that computes such a cycle from a given graph. There have been a number of systems exploiting the Curry-Howard notion of formulae-as-types. In particular we mention: Hayashi’s system PX [7], the implementation of Martin-Lof’s type theory [12], and Constable’s NuPRL [3,4]. The first two of these use logics that are not familiar to nonlogicians and the last uses its own hybrid system of logic and type theory. Our aim has always been to make the logic as close as possible to standard usage. In [6] a system of natural deduction in a very standard format is used. This system is briefly recapitulated in section 2. We build on this system. However, unlike traditional systems of mathematical logic this is a dynamic system in the sense that new axioms (or rules for induction) are constantly being added to it and in practice proofs are simplified during their construction. We work in proofEd in the same way as mathematicians: constantly introducing new functions and reusing previously proved theorems, or as computer scientists: constantly reusing (reliable) code. 2. The Logical Type Theory ( L T T ) We present a logical type theory (LTT) of many-sorted intuitionistic logic with equality (for each sort). The types are many-sorted intuitionistic formulae. The (“Curry-Howard”) terms are essentially terms in an extended typed lambda calculus that represent proofs. Reduction of Curry-Howard terms corresponds to proof normalization. The LTT is modular and extensional with respect to the operational meaning of its function terms. However the function terms may be programmed in a computational type theory. In this case we may introduce axioms for them in the LTT. These function terms can be defined in whatever way we wish, as long as they satisfy the axioms of the LTT. However the user is required to guarantee that these programs are “correct”.b Thus we retain a distinction between extensional meaning (given by the axioms they must satisfy) and intensional meaning (how they are coded in the computational type theory). aproofEd was developed from a previous system called Fred, see [5]. bThe word “correct” in this paper means “meeting its specification”.
196
Each term t E 7 has an associated sortC s - we denote this relationship in the usual fashion, by t : s. In constructing terms we shall always assume that the sort of the constructed term is appropriate. For example: If tl : s1 x s2 and t : (SI x sz + sg), then t(t1) : s3. The collection of all sorts, S, is defined as follows: We have a base collection of sorts SOthat will normally include the natural numbers, N. If s1 and s2 are sorts, then s1 + sz and s1 x s2 are sorts. We also admit SOas a sort, but SOmust not be in SO. S has associated with it a signature
Sig(S) = ({C,
: s E S } , {F,: s
E S})
where for each s E S, (i) C, is a set of function symbols (constructors) of sort s or o1x . . . xo, + s for some tuple of sorts d = o1,.. . ,on, and (ii) F, is a set of function symbols for associated functions F, : o1 x . . . x v, + s. For each sort s we also have a set of Harropd axioms A z , . The rules for first order natural deduction are readily adapted to the many-sorted case. We associate with each many-sorted formula a CurryHoward term (essentially a term of lambda calculus) representing the derivation of the rule’s conclusion. In order to normalize Curry-Howard terms we have reduction rules, see Fig. 2. We write t D u to denote that the term t reduces to the term u. Repeated application of these rules yields proof normalization. See Crossley and Shepherdson [6] for the full list of rules, all of which can be given in terms of X application and projections. The terms are formed using A, application , pairing (-, the projections fst and snd, (as usual we have the reduction rules: fst(z1,zz)= z1 and snd(zl,22) = 52) and two operations select and case that have reduction rules given in Crossley and Shepherdson [6] or Albrecht and Crossley [2]. a ) ,
=It is convenient to call the entities “sorts” rather than “types” as there are many other ‘‘types’’ in this paper. In fact for our present purposes we could easily reduce everything to first order. To do this we should just use a predicate, I n ( z , y ) , say, to represent “z is in the list y” and similarly for lists of lists. The technique is described in Henkin [8]. However we write our expressions in the conventional way and they therefore sometimes look as if they involve higher order expressions. dThe axioms one would normally employ in (constructive) mathematics are Harrop formulae (defined below). The restriction is a natural one and also has a significant effect on reducing the size of our extracted programs. Harrop axioms are axioms that are Harrop formulae and Harrop formulae are defined as follows: 1. An atomic formula or I is a Harrop formula. 2. If a: and p are Harrop, then so is a: A p. 3. If a: is a Harrop formula and y is any formula, then 7 -ia: is a Harrop formula. 4. If a: is a Harrop formula, then V z a is a Harrop formula.
197
Introduction Rules x A I- x A
Ass-I
Ax-I
when A E A x , for some sort s.
I- d f
1 df
<
>A & B
I-
I- dA F inl(d)AVB
xAI-d:B
&-I v1-I
k AxA dA+B
+-I
I- d B v2-I I- inr(d)AVB
I- d A v-I I- Ax : R d V x E R A I- dA[t:R/x:R]
)
3-1 x : R A t : R x:
Elimination Rules
xAI-dC yBI-eC I-fAVB I- c a s e ( x A , d C , y B , e C ,f A V B ) C
V-E
Figure 1. Natural deduction rules and Curry-Howard terms. Ax-I stands for Axiom introduction. Ass-I stands for assumption introduction and G is a (Curry-Howard) term variable
198
1. AX.aA-+BbA 2. AX : S.avx:S.Au : S 3. f s t ( ( a ,b) 4. snd((a,b) 5 . case(bC,xA,c c , y B , inI(a)
a[b/XIB a[u/il A[vlzl D aA D bB D b[a/xlC 6 . case(bC,zA,c c , y B , inr(a) D c[a/ylc 7. seIect(y,zP,bc, ( ~ , a ) ~ g . ~ )D b[alxlI~/Yl
Figure 2.
D D
The seven reduction rules that inductively define D.
Fig. 1 gives the natural deduction rules and the Curry-Howard terms. Note that we use I (false) and then the negation of a formula A is defined as This means there is no need for a I introduction rule as this is a A -+I. special case of +-elimination: A , A +I k 1. We make the convention that all undischarged hypotheses or assumptions are collected and listed to the left of the t- sign although we shall usually not display them. 2.1. New induction rules
Adding a sort s with constructors often gives rise to a structural induction rule in the usual manner.e This may introduce a new Curry-Howard term operation rec, with the usual fixed point semantics, and an obvious set of reduction rules. For example, in Fig. 3 we give the signature, axioms, induction rule and definition of recN for the sort of natural numbers N . An important sort for representing graphs is the parametrized list, List ( a ) ,the list of objects of sort a. The constructors of List ( a )are:
(1) E , , the empty list in List ( a ) (2) con, : a x List ( a )+ List ( a ) . We abbreviate the term con(a)(t,Z)by ( t ) :: 1. We also use ( t O , t l , as an abbreviation for the term
.....,tn)
c m , ( t o , con, (tl ,con, (...con, ( t n , E , ) ) ) ) . Intuitively a list is a higher order object but we can in fact treat lists simply as constituting a new sort. Of course we then have to ensure that eHayashi [7] has a very general rule for inductive definitions but we do not need such power for our present purposes.
199
they have the properties that we require. In this case that is easily achieved since the necessary axioms are Harrop formulae which will add nothing to the computations. Lists have the following induction rule for each sort a. Let 1 be a variable of sort List ( a )and a a variable of sort a.
This gives rise t o a recursion operator recList a with the obvious operational meaning:
recList a 6,AB us A recList ( h ):: t A B 6 Bh(recLi,t
a
tAB)
2.2. New predicates and f u n c t i o n s
An important constructive proof idiom is that of predicate definition. In ordinary mathematics, we often abbreviate a formula by a predicate. This is a useful way of encapsulating information, aids readability and helps us to identify and to use common “proof patterns”. In proof Ed, we introduce a rule of predicate abbreviation for a formula F (with zero or more occurrences of the variable z) by:
F
set P ( x )
Note that we do not allow predicates over predicates. We introduce a new function letter f of type F and the following structural meta-rule (“Template”)for any Curry-Howard term q(z) where z is a Curry-Howard term of type P : If
Template
then
set F 5 P,
fF
q( .P)Q(P)
Q(f) Q ( F )
That is, if we have formula Q that is dependent on the formula P , then we may substitute the formula F for P in Q. The converse is also a rule. Of course in doing this we must avoid all clashes of variable. Template is a means of abstracting a proof over a “formula variable”. Defining it as a structural rule is a means of avoiding higher order quantification of formula variables (as in Huet, Kahn and Paulin-Mohring [lo]) - although this could be achieved by creating a new sort (logical formulae) with a universe hierarchy (as in Martin-Lof [ 121).
200
N is a sort, representing the natural numbers.
C ( N ) = (0 : N , s : N + N } } F ( N ) = (+ : N x N -+ N} Sig(N) = < C,F > Ax(N)= { vx : N ( x = x) vx : NVy : N ( x + y = y + x ) vx : NVy : NVz : N ( x + (y + z ) = (x y) vx : N ( x + 0 = x) vx : NVy : N ( . + s(y) = s(x y))
+ +z)
+
1 Structural induction rule generated by C ( N )
--
Associated reduction rules: reCN(A)O : N fst(A)) reCN(A)S(x): N snd(d)recN(A)x Figure 3. The sort of natural numbers, the associated induction rule and the operational meaning of the recN operator.
3. The Computational Type Theory ( C T T ) Our computational type theory is the programming language ML, although it might just as easily be LISP or C++. Any language L: for which there is a mapping from terms of Church’s simple typed lambda calculus with parametrized types into C will work. We define an extraction mapping q5 from Curry-Howard terms in the LTT to terms of ML. Each sort is mapped to a corresponding ML type. For any sortf s, we assume that all the f E F, are mapped to programs for functions that satisfy the appropriate axioms Ax,. ‘Note that each parametrized sort s : Sort1 type.
+ Sort2
corresponds to a parametrized
20 1
For instance, consider the sort of natural numbers. We assume that the satisfies the axioms given in Fig. 3 for the addition function. The predefined ML function for addition will suffice, with the sort N being mapped to the ML type I n t .
ML program corresponding to
+
Theorem 3.1. Given a proof pvx:s13Y:s2a(x~Y) in the logical type theory, there is a program f an the computational type theory ML such that a ( x : s1,f ( x ) : s2) is a theorem and the extractedprogram, f = q5(p), has ML type s1- > s2 * s3 where sa is the type of the computational content of a ( x ,y ) . The proof, see Albrecht and Crossley [l] and Poernomo [13], involves defining a map, q5, from Curry-Howard terms to terms of the simply typed lambda calculus by first “deleting” computationally irrelevant CurryHoward terms: that is, by removing Harrop formulae from deductions, and then extracting the value from the first part of the Curry-Howard term.
3.1. Protocol between the CTT and the LTT Note. A fuller and formal account of this protocol may be found in our paper [13]. Just as every f E F, has a corresponding program in the CTT,every program f in the CTT has a corresponding uniqueg constant, f , in the LTT.(We assume we have an infinite number of constant function symbols
tf
i k w
.>
We have the following structural rule (Skolemization). If a ( x , y ) is a Harrop formula and t is a Curry-Howard term, then tVZ3Y4GY) ()VZOr(Z ,fa(I))
f a is the “Skolem” function. From the perspective of the associated CurryHoward terms, it means that if we have a proof t of Vls3ya(x,y), then (the universal closure of) a ( x ,f a ( y ) ) can be treated as an axiom, with f a a constant identified in the CTT with q5(t). fa is a unique function constant. In the CTT,f a is a constant representing 4(t). For example, suppose we have a proof that for all x there is a y greater than x such that y is prime: tVs3y(Prime(y)hy>z)
gThere will be no confusion caused by using the same letter as the context will make clear which is intended.
202
By Skolemization, we have the Harrop formula ()v,(Prime(f(s))Af(z)>i)
and we know that f is a unique function representing $(t) in the CTT. f and its associated Harrop formula can be used in future proofs in exactly the same way as any other function constant and its Harrop axioms (for example, just like + and the axioms for addition). For each such function with a program in the CTT we also have a reduction rule
f (5)
--)
fo
that simply implements the program for f . A related proof idiom is Function definition. This involves both the LTT and the CTT. For instance, the function length, : List ( a ) -+ N is given by the following axioms
length,(€,) = 0 Zength,((a) :: 1) = i + length,(l) These axioms define a total function length, in the LTT. We are required to specify a corresponding program in the CTT. We associate the irreflexive CTT operation of computing with the reflexive LTT equality =. The axiomatization is a (total) recursive definition, that can be automatically translated into the following M L code in the CTT: let rec length-{\alpha) = function
C I ->o
..
I a::l -> l+length-{\alpha)(l)
$ 1
Note that, in larger proofs when we are anxious t o reduce the size of the term (program), we may choose to implement the associated program in a manner different from that suggested by the axiomatization. This is an important feature of our approach - intensionally distinct programs in the CTT correspond to extensionally interchangeable functions in the LTT. Of course, the programs extracted from our system are only as correct with respect to the axiomatization as these programs are correct (and correctly, though usually trivially, axiomatized) . As noted above, axiomatizations of functions in the LTT and their associated computational definitions in the CTT are separate. In many con-
203
structive proofs, functions are not proved and extracted: instead, a total function is defined by an axiomatization. 4. Representing graphs in the formal system
We consider a standard axiomatization of the theory of graphs, G, in terms of vertices and edges. The vertices will be represented by positive integers. Consider the graph with four vertices in Fig. 4a represented by the four element list of lists of neighbours ((1,2,3),(2,1,3),(3,1,2), (4))where each element is of sort List(N). Not all lists of elements of sort List(N) corre-
Figure 4.
Two sample graphs
spond t o graphs: in a graph the edge relation is irreflexive and symmetric. The list above has the properties (1) The nth member of the list is a list of numbers beginning with n. (2) (Symmetry)If the nth member of the list is a list containing m and m # n, then the mth member of the list is a list containing n. (3) Each member of the list is a repetition-free list of numbers.h
These properties are expressible in our formal system for G with the aid of certain extra function symbols, that we now define. Note that each function is provably total in the formal system. Here is the list of required functions in F ~ i ~and t , the associated axioms. All formulae are considered to be universally closed. We note that appropriate M L definitions can be generated automatically as in the previous section. (1) A binary function memberN of two arguments: a natural number, n, and a list.' The function computes the n th member of the list. hThis ensures that the edge relation is irreflexive and that no pair of vertices are joined by more than one edge (viz. the graph is a simple graph). 'For lists of elements of sort a we use member, as the function letter.
204
Since all functions are total we will need to use a “default va1ue”j for cases where n is larger than the length of the list or where n = 0. The definitions for the cases a = N , List(N) are given below. In all cases m is a variable of sort N , and 1 is a list variable of sort List(a),and a is a variable of sort a. The last four items are defined by list recursion.
memberN(0,Z)= 0 memberN(m,E N ) = 0 memberN(1,( a ) :: 1 ) = a memberN(m 1 , (a) :: 1 ) = memberN(m,1) memberLi,t(N)(0,O= E N memberList(N)(m,E L i s t ( N ) ) = EN memberList(N)(l, ( a ) :: 1) = a memberList(N) (m+ 1, ( a ) :: 1) = memberList(N) (m,1 )
+
(2) List successor, S. This function takes a list as argument, adds one to each number in the list and returns the revised list.
S(E) = E S((a) :: 1) = ( a + 1) :: S(1) (3) Position function, Zistpos. listpos(n,1) gives a list of all the positions the number n takes in the list 1. If the list 1 does not contain n then the empty list is returned. We take the head position as 0, so position Ic corresponds to the k + lst member of the list. Zistpos~n,E ) = € listpos(n, ( a ) :: 1 ) = (0) :: S(listpos(n, I)) if n = a Zistpos(n, ( a ) :: 1) = S(Zistpos(n,1)) if n # a
(4) Initial segment of a list, initlist. initlist(k,Z) computes the list consisting of the first Ic + 1 elements of the list 1, if Ic 1 > length(2) then the list 1 is returned.
+
initlist(k,E ) = E initZist(0,( a ) :: 1 ) = ( a ) initZist(Ic+ 1,( a ) :: 1) = ( a ) :: initZist(Ic,1) jNote that the default value for the first case below is 0. Because all our graphs contain only positive integers, it is always the case that when we apply our functions to lists of vertices we shall be able to decide whether we are getting a vertex or the default value.
205
(5) Tail segment of a list. We define a function tail (1, n) that has a list 1 (of natural numbers) and a number n as arguments and computes the list obtained by deleting the first n members of 1.
tail (El n) = € tail (1,O) = 1 tail ( ( u ) :: 1, n + 1) = tail (1, n)
5. Cycles in even parity graphs Once all the functions above are defined in proof Ed, we can set a predicate gruph(1) to mean that a list I of sort List(List(N))represents a graph.k The formula gruph(1) is defined in proofEd by the conjunction of four Harrop formulae:
set gruph(1) length(1) 5 1 + IA Vi(1 5 i 5 Zength(Z) + memberlv(l,’member~ist(~~(i,Z)) = i) A Vi (1 5 i 5 length(1)-+ rep~ree(memberList(lv)(i, 1))) A ViVj(((1 5 i 5 length(l)A (1 5 j 5 length(l)A j # i)) -+ listpos(j, member(i,1 ) ) # E + Zistpos(i, member(j,1 ) ) # 6)) where repfree(1)is a predicate (meaning “free of repetitions”) defined by
set repfree(1)= Vn((length(listpos(n,1 ) )
> 1) + I)
A graph has even parity if the number of vertices adjacent to each vertex is even. So each list in 1 must have an odd number length. Consider the function from lists of numbers to numbers defined by par(€)= 0 pur((u) :: 1 ) = 1 - par(1) where is the “monus” function.’ Then I is a list describing an even parity graph if evenpar(1):
set evenpar(1) graph(1) A V i ( 1 5 i 5 Zength(Z)+ pur(member(i,1 ) ) = 1) To motivate our method for cycle detection look again at the list 1 corresponding to the graph of Fig. 4a, with the given adjacency matrix above, kWe exclude trivial graphs consisting of one or zero vertices.
’Monus: x
1y is defined by I
y = x - y if
I
2 y and = 0 otherwise.
206
((1,2,3),(2,1,3),(3,1,2),(4)).Note that the same graph is represented by taking the first member as (1,3,2),the order of the numbers in the tail of each of the elements in the list 1 is not important. Now to locate a cycle we start by locating the first element in 1 that is a list of length > 1. This is (1,2,3) so we begin tracing a path with vertex 1 and since the first vertex mentioned in this list after 1 is vertex 2 we choose the edge from 1 to 2. Now scan the (tail of) the list (2,1,3) in 1 corresponding to vertex 2 for the first vertex not equal to 1 (we do not leave 2 by the same edge we arrived), this gives vertex 3 and so we now scan the (tail of) list (3,1,2) for the first vertex not equal t o 2. This leads to 1 and then 2, etc. Continuing in this manner we can construct a list of adjacent vertices ( 1 , 2 , 3 , 1 ,...) of arbitrary length. Such a list defines a walk in the graph. In proofEd, for c, a list of numbers (vertices), and 1 E List(List(N)),a graph (viz. a list of lists of numbers), we set waZk(1,c) as an abbreviation for the Harrop formula that is a conjunction of four formulae
set waZk(1,c) = length(c) > 1 A gruph(1) A Vk(1 5 k < length(c) + listpos( member( (k + 1),c ) , memberLi,,(N) (member(k,c ) ,1 ) ) # E ) A Vk(1 < k < length(c) + member(k + 1, c ) # member(k - 1, c)) The first occurrence of a repeated vertex yields a cycle represented by the sublist of the vertices between the repeated vertices, in this case ( 1 , 2 , 3 , 1 ) . Note that the desired sublist does not necessarily begin a t the vertex we start from, although in this case it does happen that way. To carry this construction over to the formal system we need a function that searches a list for the first element not equal t o a given number. The function spotadiff is defined so that spotadiff(1,m) gives the first element in the list 1 that is not equal to m; if there is no such element then the default value, 0, is returned. It is given by axioms:
spotadiff(€,m) = 0 a # m + spotadiff((a) :: 1,m) = a a = m + spotadiff((a):: 1,m) = spotadiff(1,m)
As usual it may be programmed independently in the CTT - by our convention we are simply required to guarantee that it satisfies the axioms. To start the construction we also need a function, which we call start, that takes as its argument a list, 1, of lists of numbers and returns the head of the first list in 1 that has length greater than 1, i.e. locates the first
207
non-isolated vertex. If there is no list in 1 with length 0 is returned.
> 1 then the default
start(€)= 0 length(a)> 1 + start((a) :: 1) = member(1,a) length(a)5 1 + start((a) :: 1 ) = start(1) As usual, the function symbol start corresponds to a program in the CTT that satisfies the axioms. Finally, the function gen that generates a list of adjacent vertices from the list 1 specifying the graph can now be defined. gen(1,n ) gives the vertex for the nth stage of construction. It has the following axioms
gen(1,O)= start(1) gen(1,1) = member~(2, memberLUt(N)(start(Z), 1) gen(1,n 2) = spotadiff(taiZ (memberList(N)(gen(l,n l),Z)), l ) ,gen(1,n))
+
+
If 1 is a list corresponding to an even-parity graph then the function gen(1,m)is either identically zero (in the case that 1 has no edges) or the function is never zero and gen(1,rn) and gen(1,m + 1) are adjacent vertices for every m. We need to make sure that we have a term in our language to represent a list of the form (gen(Z,O),gen(1,l),. . . , gen(1,k)) for any k,1. Actually it is easier to define a function that computes the reverse of this list. We define a new function genZist(1,k)where 1 is a term of sort List(List(N)), and k is a term of sort N . The function has values of sort List(N) genlist(1,0 ) = (gen(1,0 ) ) genZist(Z,k + 1) = (gen(1,k
+ 1 ) ) :: genlist(Z,k)
So genZist(1,k) corresponds to (gen(1,k), ...g en(1,0 ) ) . 6. The Proof
In proofEd, just as in mathematics practised by mathematicians, we can build up a proof in layers, using earlier layers in order to achieve the next layer. In this section we examine the topmost layer, where the required theorem is proved using several lemmas which we assume we have already proved. If c is a list of numbers then cycle is a predicate defined in proofEd: set cycZe(c,1)
member(1,c ) = member(length(c),c) A repfree(tai1(c,1))A walk(1, c )
208
The Main Theorem we want to prove is V1 (evenpar(1) A start(1) # 0 + 3c( cycle(c, 1 ) ) )
This says that if Z represents a graph that does not consist entirely of isolated vertices, then Z contains a (non-trivial) cycle. We let the predicate genZistGivesWalk(Z) stand for the statement that the function genlist generates walks in the graph 1; from these walks we wish to extract a cycle: set genlistGives Walk(1) evenpar(1)A start(1) # 0 + Vm ( m > 0 + walk(l,genlist(1,m))) This can be proved by observing that a cycle in a graph(represented by a list 1) can be represented by a list c such that (i) each pair of successive members defines an edge of the graph (corresponding to Z), (ii) the first and last entries of the list c are the same and (iii) these are the only repeated vertices in c. However, note that genZistGivesWaZk(1) is represented by a Harrop formula - it has no constructive content. Because it is Harrop it does not contribute to the computation and nor does its proof. It therefore does not matter whether we establish this constructively or even classically. We can just take it as a new (computational-content-free) axiom. The proof of the Main Theorem relies on the lemma following that states that it is provable that any list of numbers is either repetition free or the list contains an element (say a) such that for some tail segment of the list the element a occurs exactly twice in the segment and no other element occurs more than once in this tail segment.
Vl(repfree@)V ListHas UniqueEltOccursTwiceInTail ( 1 ) ) where ListHas UniqueEltOccursTwiceInTail (1) is a predicate defined by: set ListHas UniqueEltOccursTwiceInTaiZ( I ) l a 3k 3m (listpos(a, tail (1, k)) = (0, m
+ 1A
repfree(tad (1, k
+ 1)))
This Main Lemma is proved in Appendix A. The constructive proof of this lemma is the key to cycle extraction. Note that the proof of the lemma involves understanding what the predicate ListHas UniqueEltOccursTwiceInTailstands for. However, once we have proved the lemma, the definition of this predicate can be “encapsulated” and not looked at again. We do not need to look at the definition now to prove the theorem. The M L program extracted for the lemma is disp!ayed in Fig. 5 where KSC158 and Cgr20 are programs corresponding to other lemmas used in the proof of the lemma - see Appendix C for a full listing of their programs.
209
In this program we can see how recursion corresponds to induction. Here the base case of the recursion occurs in lines 2-4 and the rest of the program corresponds to the induction step, while the actual recursive call comes in the function call for fun81 four lines from the end. Inside the induction step the most important calls are to Cgr20 (see Appendix C.3) and KC158 (see Appendix C.4) below. The first step towards the Main Theorem is to use V elimination on the lemma, replacing Z by the term genlist(1, length(1) 1).
+
+
VZ(repfree(genZist(1,length@) 1))v ListHas UniqueEltOccurs TwiceInTail (genlist(1,length(1) + 1))) We can establish the deduction (see Appendix B) of
evenpar(Z) I- repfree(genZist(l,length(Z) + 1))+ I Note however that this formula is Harrop - so has no constructive content, and is of course true in the intended model. (If 1 only contains numbers 1 , . . . ,n, then a list of length n 1 constructed from Z must have a repetition). We can take therefore take this formula as an axiom. This Harrop axiom together with the formula obtained from the lemma above by V-elimination can be substituted into the following proof pattern (using applications of I-elimination and V-elimination):
+
AVB A,A+II-B B
BI-B
This gives us
ewenpar(Z) t- ListHas UniqueEltOccursTwiceInTail(genZist(Z,Zength(Z) + repfree( tail (genZist(1, length(1) l),k
+
and therefore (1)
evenpar(1) I- ListHas UniqueEltOccursTwiceInTail (genlist(Z,length(Z) + We can also obtain (2)
start(1) # 0 A evenpar(l), ListHas UniqueEZtOccursTwiceIn Tail (genlist(1,Zength(1)
+ 1))
I- 3c (cycle(c, 1 ) )
(1) and (2) give our theorem.
210
let Cgr21 = let rec fun80 1 = begin match 1 with [ 3 -> inl(1et fun100 x = (s 0 ) in fun1001 I h::t -> let fun81 z = let fun82 1 = let fun83 X217 = begin match (X217) with inl(g) -> ((let fun92 X218 = begin match ( ( ((Cgr20 1) X218) z)) with inl(g) -> (inl(1et fun98 x = begin match (((KSC158 x) z ) ) with inl(g) -> ( 0 ) I inr(g) -> (X218 x) end in fun98)) I inr(g) -> ((let fun93 X221 = inr ( (z (0 ((X221) let fun94 x = (app X221 (let fun95 y = (X218 x) in fun95)) in fun94)))) in fun93) g) end in fun92) g) I inr(g) -> ((let fun84 X219 = (select X219 (let fun85 b = let fun86 X241 = inr((b (select X241 (let fun87 c = let fun88 X242 = ( ( s c ) ((pi1 X242), let fun89 x = ( (pi2 X242) x) in fun89)) in fun88 in fun87)))) in fun86 in fun85)) in fun84) g) end in fun83 in fun82 in fun81 h t (fun80 t) end in fun80 J
J
J
.. J J
Figure 5 . ML program (for Cgrll) extracted from the proof of the Main Lemma: V1( repfree(1) V ListHasUniqueEltOccursTwiceInTail( I ) )
21 1
(1) For the proof of (2) the witness for c is the initial segment of the list
tail (genlist(l,length(Z)
+ l),k )
consisting of elements in positions 0 t o m inclusive instantiated to the term
-
so c will be
initZist(m,tail (genZist(l,length(Z) + l),k)). To establish (2) we use the true Harrop formula (3) (see next comment) :
VZ V k Va Qm ((sturt(l)# 0 A ewenpar(l))A (listpos(a,tail (genlist(l,length(l) l), k)) = (0,m)A repfree(tai2 (genlist(l,Zength(Z) I),k I)))) -+ cycle(initlist(rn,tail (genlist(l,Zength(l) l),k)),1 ) )
+
+
+
+
jFrom this we form the deduction
((start(l)# 0 A ewenpar(2)) + (listpos(a,tail (genlist(l,length(Z) l),k)) = (0, m 1) + repfree(tail (genlist(l,Zength(Z) l),k 1)))) t- cycle(initlist(m, tail (genlist(l,length(l) l),k)),1)
+
+
+
+
+
3-introduction gives:
((start(l)# 0 A ewenpur(Z))A (Zistpos(a,tail (genlist(Z,Zength(Z) I),k)) = (0, m 1)A repfree(tuiZ (genZist(l,length(l) l),k t- 3c (cycle(c,1 ) )
+
+
+
+ 1))))
Application of 3-elimination (3 times) finally gives
sturt(l) # 0 A ewenpur(l) A 3a 3k 3m ((Zistpos(a,tail (genZist(Z,Zength(l) l),k))(O, m + 1) A repfree(tad (genlist(l,kngth(l) + I),k + 1)))) t- 3c cycle(c,l)
+
By our definition of ListHas UniqueEltOccursTwiceInTuil(1), this is equivalent to (2). (2) We shorten the proof by not giving a formal proof of the formula (3). This is a Harrop formula and therefore has no computational content, therefore, since it is true (in the intended model), we can take it as a new axiom.
212
In establishing (2) above it may look as if we are cheating and simply stating that the list for the cycle is initlist(m,tail (genlist(l, length(1) l),k)). In fact the computational content of this is all in the proof of the Main Lemma. This proof yields an algorithm that, given a list, extracts a sublist with the property that the first and last elements are equal, and that there are no other repetitions in the sublist. We are applying this algorithm to a particular list generated from the graph list 1 via genlist. We “trust” genlist to generate a walk from list 1 and then apply our constructive proof of the Main Lemma to this generated list. Suppose we were to use V-elimination with t on the theorem for the graph we wish to use. Then (provided we are in fact dealing with a term t that represents a non-trivial even parity graph), we could add the Harrop axiom evenpar(t) A start(t) # 0 to obtain a proof of 3c(cycle(c, t ) ) . This proof will normalize to give a term for c that represents the cycle. The final program uses the program Cgr21 for the Main Lemma (see Fig. 5) and is as follows:
+
l e t main = l e t fun96 1 X = begin match ((Cgr21 ( g e n l i s t 1 (s i n l ( g ) -> C 1 I i n r ( g ) -> ( ( l e t ( s e l e c t (X40) ( l e t fun98 b = l e t fun99 X43 = ( s e l e c t (X43) ( l e t fun100 c = l e t fun101 X44 = (app ( ( p i 1 X44)) ( l e t fun102 y ( i n i t l i s t (y+l) ( t a i l (genlist i n funl02)) i n fun101 i n funl00)) i n fun97)g) end i n fun96
(length 1) 1) ) ) with fun97 X40
=
= 1 ((length 1)+1) ) c ) )
i n fun99 i n fun98))
2 2
Here the M L items app and s e l e c t are aliases for function application defined by l e t app x y = (x y);; l e t s e l e c t x y = (y x);;
21 3
Note that the main function takes an input 1 for the graph we want to use and also an input X. X should stand for a term mapped by the extraction map, 4, from a proof that evenpar(t) A start(t) # 0. However, that statement is Harrop, so X can be anything (because it is not used in the computation). This is somewhat unsatisfactory, although not unexpected: it follows from comment 4. So main is correct modulo whether ewenpar(t) A start(t) # 0 is true or not. If we go on to prove
eerenpar(t) A start(t) # 0 V T(evenpar(t) A sturt(t) # 0) then we can extract a program to determine if ewenpar(t)Astart(t) # 0 is true or not, and then use this to extract a program defined for all graphs that calls main only if ewenpar(t) A start(t) # 0 is true, and returns some “error” value if not. As a further refinement it is also possible to create a new “predicate subtype” (see Rushby, Owre and Shankar [14]) T 5 graph of evenpar(t) A start(t), and alter the map graphs, such that t : T 4 so that main is defined only for t : T .
7. Demonstration results Finally we present some practical results. Here is the result for the graph with four vertices in Fig. 4a. #main [[1;2;31;[2;1;31;C3;1;21;[411;; - : i n t l i s t = [I; 3; 2; I1 Next we consider the even parity graph in Fig. 4b with vertices 1,.. . , 6 and extract a cycle in it. [4;3;51;C5;4;31;[6;1;311;; #main [[1;2;6];[2;1;31;[3;2;4;5;61; - : i n t l i s t = C3; 5; 4; 31 8. Conclusion
We have demonstrated a system for extracting programs from proofs in a very natural (first-order) logic which allows us directly to use programs that we have previously constructed. In our example we constructed a program Cgr2i which we had previously extracted from a proof, and then we used that program, called as Cgr21 in our main program. Thus we are able to build on our earlier programs directly in our logic. The system has been
214
demonstrated in an example from graph theory and, because of our layering of programs and proofs, the final programs are, within limits, readable by humans. In fact it is not as readable as we might like because we have performed certain optimizations, in particular reductions involving Harrop formulae. We have therefore a balance between legibility and optimal coding. However our main program is very short not only because it calls previous programs that we have either extracted from proofs but also because we are able t o use programs from the standard programming language ML. In fact, in the case of our main program the abstract structure of the proof is fairly clearly reflected in the M L program we extract. This leads to the program being modular over the previously extracted programs. Modularity is an important issue in software engineering and our method represents a step in formalizing modular construction of programs. References 1. David Albrecht and John N. Crossley. Program extraction, simplified proofterms and realizability. Technical Report 96/275, Department of Computer Science, Monash University, Australia, 3800, 1996. 2. David William Albrecht and John Newsome Crossley. Program extraction, simplified proof-terms and realizability. Technical Report 271 , Department of Computer Science, Monash University, Australia, 3800, 1997. 3. Robert L. Constable, Stuart F. Allen, H. M. Bromley, W. R. Cleaveland, J. F. Cremer, R. W. Harper, Douglas J. Howe, T. B. Knoblock, N. P. Mendler, P. Panangaden, James T. Sasaki, and Scott F. Smith. Implementing Mathematics with the Nuprl Development System. Prentice-Hall, NJ, 1986. 4. Robert L. Constable, Stuart F. Allen, H. M. Bromley, W. R. Cleaveland, J. F. Cremer, R. W. Harper, Douglas J. Howe, T. B. Knoblock, N. P. Mendler, P. Panangaden, James T. Sasaki, and Scott F. Smith. Implementing Mathematics with the Nuprl Development System. Prentice-Hall, NJ, 1986. 5. John Newsome Crossley and Iman Poernomo. Fred: An approach to generating real, correct, reusable programs from proofs. Journal of Universal Computer Science, 7:71-88, 2001. 6. John Newsome Crossley and John Cedric Shepherdson. Extracting programs from proofs by an extension of the curry-howard process. In John Newsome Crossley, Jeffrey B. Remmel, Richard A. Shore, and Moss E. Sweedler, editors, Logical Methods, pages 222-288. Birkhauser, Boston, MA, 1993. 7. Susumu Hayashi and Hiroshi Nakano. PX - A Computational Logic. MIT Press, Cambridge, MA, 1988. 8. Leon Henkin. Completeness in the Theory of Types. Journal of Symbolic Logic, 15:81-91, 1950. 9. William Howard. The formulae-as-types notion of construction. In John Roger Hindley and Jonathan Seldin, editors, To H.B. Curry: Essays
215
10.
11.
12. 13.
14.
on Combinatory Logic, Lambda Calculus, and Formalism, pages 479-490. Academic Press, 1969. Gerard Huet, Gilles Kahn, and Christine Paulin-Mohring. The Coq Proof assistant Reference Manual: Version 6.1. Inria, Coq project research report RT-0203 edition, 1997. Georg Kreisel. Interpretation of analysis by means of constructive functionals of finite types. In Arend Heyting, editor, Constructiuity in Mathematics, Proceedings of the Colloquim held at Amsterdam in 1957, pages 101-128. North-Holland, Amsterdam, 1959. Per Martin-Lof. Intuitionistic Type Theory. Bibliopolis, Naples, Italy, 1984. Iman Poernomo and John Newsome Crossley. Protocols between programs and proofs. In Kung-Kiu Lau, editor, Logic Based Program Synthesis and Transformation, 10th International Workshop, L O P S T R 2000 London, UK, July 24-28, 2000, Selected Papers, volume 2042 of Lecture Notes i n Computer Science, pages 18-37. Springer, 2001. John Rushby, Sam Owre, and N. Shankar. Subtypes for specifications: Predicate subtypes in PVS. IEEE Transactions on Software Engineering, 24(9):709-720, 1998.
APPENDIX A We establish the Main Lemma
V1(repfree( 1) V 3a3k3m(Zistpos(a, tad (1, k)) = (0, m
+ 1) A
repfree(tazl(1, k + 1))))
by list induction. We introduce A(a,Ic, m, 1) using Template:
set A ( a , k , m , l ) listpos(a,tail(l,k))= ( O , m + l ) A r e p f r e e ( t a z Z ( l , k + l ) ) Base case. 1 = E N In this case we have repfree( E )
so by V-introduction we obtain repfree(€)V 3a3k3mA(a, Ic, m, Z)
Induction step. We have to show VbVZ((repfree(l)V 3a3Ic3mA(a,k,m,Z))+ (repfree((b) :: 1) V 3a3k3mA(a7k , m , (b) :: I ) ) )
216
It suffices to obtain the deduction: repfree(1) V 3a3k3rnA(a,k,rn,I ) I- repfree(@) :: I) V 3a3k3rnA(a,k,rn, ( b ) :: 1)
(2)
since an application of +-introduction followed by two applications of V-introduction gives the induction step. To establish (1) we show repfree(1) I- repfree((b) :: 1) V 3a3k3rnA(a,k,rn, (b) :: I )
(2)
and
3a3k3rn(A(a,k,rn,I ) I- repfree((b) :: 1) V 3a3k 3rn(A(a,k,rn, (b) :: I )
(3)
We begin with (2). We first establish the deduction: repfree(I),listpos(b,Z) = E V 3r(Zistpos(b,I ) = ( r ) ) I- repfree(@) :: I)
v 3a3k3rnA(a,k,rn, ( b ) :: 1 ) ) (4)
Then since (from Appendix B) we have the lemma: repfree(2) I- Zistpos(b, 1) = E V 3rZistpos(b,1) = ( r )
(5)
Now for (4),first note that we can obtain (see Appendix B) repfree(I),Zistpos(b, 1) = E I- repfree(@) :: I ) so by V introduction we obtain the expression on the right of the logical Iin (4).We can also show (see Appendix B)
Iistpos(b,I ) = ( r ) I- Iistpos((b) :: I ) = (0, r
+ 1)
and then repfree(I) A 3rZistpos(b, I) = ( r ) )I- 3a3k3rnA(a,k,rn, (b) :: I )
+
where the witnesses for a, k,m are b, 0, r 1respectively. Intuitively, we are saying that if we have a repetition free list (c, ..,b, ...) then adding b to the head gives (b,c, ...,b, ..) and b is the only repeated entry. An application of V introduction to this deduction then gives the required conclusion, repfree((b) :: 1) v 3a3k3rnA(a,k,rn, (b) :: 1)
This establishes (4)and hence also (2). We now establish (3), by showing:
3a3k3rnA(a,k,rn,1) I- 3a3k3rnA(a,k,rn, ( b ) :: 1 ) )
217
Then (3) will follow by an v introduction. Recall that A(a,k,m, 1) is
listpos(a, tail (1, k)) = (0, m
+ 1) A repfree(tail (I,k + 1))
Using the definition of tail we can easily establish
tail (1, k) = tail ( ( b ) :: 1, k
+ 1)
Hence we have listpos(a, tail (2, k)) = lZstpos(a, tail ( ( b ) :: 1, k + 1))
So we have the deduction
+
+
listpos(a, tail (1, k)) = (0, m 1)A repfree(tail (1, k 1)) I- listpos(a, tail ( ( b ) :: 1, k + 1)) = (0,m 1) A repfree(tail ((b) :: 1, k + 2))
+
3 introduction applied three times followed by 3 elimination also applied three times gives 3a3k3mA(a,k,m, 1) I- 3a3k3mA(a,k,m, (b) :: 1 ) which establishes ( 6 ) , hence (3), and the induction step is finished.
APPENDIX B 1. We establish the lemma used in Appendix A, repfree(1) I- listpos(b, 1) = E
v 3r listpos(b, 1 ) = ( r )
that is, VnZength(listpos(n,1 ) )
5 1 I- listpos(b, 1) = E v 3r listpos(b,1) = ( r )
The following theorems are easily established by list induction Vl(Zength(1) = 0
+ 1 = E)
W(length(1) = 1 + 3r 1 = ( r ) )
and then an application of V elimination replacing 1 by listpos(b, 1) gives the result. 2. We establish repfree(l), Zistpos(b, 1) = E t- repfree((b) :: 1 )
218
that is,
V n Zength(Zistpos(n,1 ) ) 5 l,Zzstpos(b,1) = E t- V n Zength(Zistpos(n, (b) :: 1 ) ) 5 1 This can be established by showing
Iength(Iistpos(n, 1 ) ) 5 1,n # b t- Zength(Izstpos(n,(b) :: I ) ) 5 1 and
length(Zistpos(n,I ) )
4 1,n = b, lzstpos(b,1) = E t- Zength(Zistpos(n,(b) :: 1 ) ) = 1
3. The theorem
Zistpos(b,Z) = ( T ) -+ listpos(b, ( b ) :: I ) = (0,r
+ 1)
follows easily from the definition of Zzstpos. 4. The proof of the Main Theorem used the deduction
evenpar(Z,n ) t- repfree(genZist(Z,n + 1))-+ I where we define evenpar by overloading evenpar as the two place function defined by
set evenpar(2,n)
evenpar(Z) A Zength(Z) = n.
We now establish this result. We need to introduce a new binary function sum that has as arguments a list of natural numbers and a natural number. The definition is
sum(I,0 ) = Zength(Zistpos(0,Z)) sum(Z,k
+ 1) = sum(Z,k ) + Zength(Zzstpos(k + 1,Z)) i=k
So sum(I,k) computes
C Zength(lzstpos(i,1))
i=O
The next two lemmas are established by induction on n. Lemma A. V n ( b 5 n -+ surn((b):: 1, n ) = 1 surn(Z,n ) ) .
+ Lemma B. Vn(repfree(E) -+ surn(Z,n) 5 n + 1). Note that
Z is a free list variable here, and b is a free number variable.
Lemma C. VZ(Vi member(i,I ) 5 n -+ surn(Z,n ) = Zength(Z)). This is established by list induction.
219
Base case. 1 = E . We can easily show s u m ( ~n) , = length(€)
and the result follows. Induction step. It suffices to show V i member(i,Z) 5 n
-+sum(1,n) = Zength(Z),
V i member(i, (b) :: 1 5 n) I- sum((b) :: 1,n) = Zength((b) :: 1 )
and this follows from Lemma A since the hypothesis b 5 n is implied by V i membedi, (b) :: 1 ) 5 n so that sum(@) :: I , n) = 1 sum(1,n) = Zength((b) :: 1 ) follows from the hypotheses for the induction step.
+
At last we can show the result we are seeking
evenpar(1,n ) I- repfree(genZist(1,n
+ 1))+ 1
The definition of evenpar(1,n) allows us to show
evenpar(2,n) k V i member(i, genZist(1, n + 1))5 n
+ 1) replacing I to obtain evenpar(1,n ) I- surn(genZist(1,n + 1))= Zength(genZist(1,n + 1))
now apply Lemma C with genZist(1,n
Now Lemma B gives
repfree(genZzst(Z,n + 1))+ sum(genZist(Z, n
+ 1)) 5 n + 1
+ 1))= n + 2, we have evenpar(1,n ) ,repfree(genZist(2,n + 1))k 1
but since k Zength(genZist(1,n
and we are done.
APPENDIX C Here is the listing of the M L functions called in the program for Cgr21. Each function is generated with an accompanying “documentation” - the formula t o whose proof the function corresponds which is given in square brackets as a comment to the program.
220
1. The program for KSC137b
[ALL x . CCx=Ol I CCx=s(O)l I [EXISTS a . Cx=s(s(a>>1]111 let KSC137b = let rec fun65 x = begin match x with 0 -> inl(unit) I _ -> let fun66 x = let fun67 X46 = inr ((let rec fun68 x = begin match x with 0 -> inl(unit) 1- -> let fun69 x = let fun70 X47 = inr(x) in fun70 in fun69 (x-1) (fun68 (x-1)) end in fun68 x)) in fun67 in fun66 (x-I) (fun65 (x-I)) end in fun65 (*
*>
2. The program for KSCl33
.
[ALL y-natnum . [ [(x"natnum+y"natnum)=s(O)l--> C[x-natnum=s(O>l I ~y~natnum=s(O)lllll*) let KSCl33 = let fun71 x = let fun72 y = begin match ((KSCl37b x)) with in1(g) -> (inr(unit) ) I inr(g) -> ((let fun73 X40 = begin match (X40) with inl(g) -> (inl(unit)) I inr(g) -> ((let fun74 X41 = (select X41 (let fun75 a = unit in fun75)) in fun741 g) end in fun73) g) end in fun72 in fun71 (* [ALL x-natnum
.. S
l
22 1
3. The program for Cgr20
[ALL 1-List . [[ALL x-natnum . [EXISTS a-natnum . ~(fd(listpos~x^natnum,1^List>>+s~a~natnum>>=s(s(O>~l11--~ [ALL x-natnum . CClistpos(x^natnumJ1~List)=ernptlist^Listl I [EXISTS y-natnum . [listpos(x^natnum,l^List)= cons(y”natnum,kemptyseq^List>llllll *> let Cgr20 = let fun76 1 = let fun77 X31 = let fq78 x = begin match ((select (X31 x) (let fun79 a = a> unit>> with begin match ((((KSC133 (fd (listpos x 1) (*
>>
inl(g) -> (inr(unit>> I inr(g> -> ((inl(unit>) end in fun79))) with in1(g) -> (in1(unit> ) I inr(g> -> ((inr(((Cgrl9 (listpos x 1)) unit>>> end in fun78 in fun77 in fun76 J J
4. The program for KSCl58 (*
[ALL x
. [ALL y . [[x=yl
I ~Cx=yl--~Bottom1111 *>
let KSC158 = let fun54 x = let fun55 y = begin match (((KSCl39 x> y>> with inl(g) -> ((let fun59 X45 = inr(unit) in fun59)g) I inr(g> -> ((let fun56 X52 = begin match (X52) with inl(g) -> (inl(unit>> I inr(g> -> ((let fun57 X54 = (select X54 (let fun58 a = inr(unit) in fun58)) in fun57)g) end in fun56)g) end in fun55 in fun54
..
2 3
222
A COMMON STRUCTURE OF LOGICAL AND ALGEBRAIC ALGORITHMS
KAWAGUCHI, YUUICHI Dept. of Liberal Arts and Sciences, Tenshi College 31 -2, Kita 13, Higashi 3, Higashi-ku, Sapporo, Hokkaido 065-0013 Japan E-mail: yuuichiOtenshi. a c . j p In this paper, it is shown that there is a common structure of algebraic algorithms and logical algorithms. Three examples of problem-solving are shown. Simultaneous equations for describing and solving the problem are used in one example, and congruence expressions are used in another example. Both of these problem are algebraic. Logical formulae and the resolution principle are used in the third problem, which is a logical problem. The three examples are formalized by using three basic concepts, a description of a given problem, an answer to the problem, and the relationship between these two. In the formalization, the algorithm always consists of a sequence of transformations of descriptions. When a description is transformed into another form, the algorithm is guaranteed to be correct, i.e., the correct answer is obtained, if the transformation keeps the answer not changed.
K e y Words: problem solving, program transformation, common structure.
1. Introduction 1.l. Common Structure
It has been shown that there is a common structure of algorithms for solving algebraic problems [6]. In this paper, it is shown that algorithms for solving a logical problem have the similar structure. Note that the meaning of the word ‘structure’ in this paper is different from the one in logic [9]. Three examples of problem-solving are shown. Two of the problems are algebraic, and one is logical. In this paper, it is shown that an equation holds in the three examples and that the algorithm used for solving the problem always consists of a sequence of transformations satisfying that equation. All of the examples shown in this paper have already been solved. Both of the algebraic problems have efficient method for solving. The logical
223
problem also have a method for solving, but it is not efficient. However, the existance of a common structure suggests that there is also an efficient method for solving for the logical problem. In order to solve a given problem, the problem must be described in a formal system. The answer is bound to the description by a certain relationship. In formal systems, relationships are expressed by using maps. Let D be a set of all formal descriptions and A be a set of all answers. Given that the formal description of a given problem is d E D ,the correct answer to it is a E A , and the map that expresses the relationship is f : D + A, then it is shown that a = f(d) holds in all three examples. Note that an element in the set A is not always correct. The set describes only the shape of each answer. The word ‘formal system’ is different from that in logic [9]. It is a general one. The ‘formal description’ is not limited to only logical formulae. In general, it is difficult to compute f ( d ) directly. Suppose that there is another description, d‘, that satisfies a = f (d’) and that makes computation off (d’) easier. It is reasonable to use f (d’) for obtaining the correct answer a. In this case, the algorithm for solving the problem is one by which the original description is transformed into the description d’. If d can not be transformed into d’ in one step, then the algorithm is a sequence of transformations. Suppose that a sequence d = d l , d2,. . . ,d, = d‘ are made by the algorithm, where di is transformed into di+l for each i = 1,2, . . . ,n - 1. If it holds that a = f ( d ) = f(d2) = ’ . * = f ( d n ) , then the algorithm is guaranteed to be correct, i.e., the answer obtained is correct. A transformation that fulfills this condition is called an ‘equivalent transformation.’ It is shown that each of the algorithms used to solve the three examples consists of equivalent transformations.
1.2. Related Work The idea for this paper originates from Akama’s work [I]. The work proposed a computational framework based on equivalent transformations. Computation in the framework is guaranteed to be correct and is more sufficient than that in the logic programming paradigm. An alternative transformation method for logical formulae, folding and unfolding is described in a book [3]. There are many books and papers on formal description of problems and automatic generation of algorithms (e.g., Dijkstra [2] and Kowalski [7]). The focus of this paper is on common structure of algorithms.
224
2. Three Examples 2.1. Cranes and Tortoises
Let us consider the following problem. Suppose that there are some cranes and tortoises. The total number of heads of the cranes and tortoises is 35, and the total number of legs is 94. Given this information, how many cranes and tortoises are there? This problem can be expressed by the following simultaneous equations: x+y=35
2a:
+ 4y = 94
The symbol a: is the number of cranes and y is the number of tortoises. The correct answer to the problem is the pair of integers x and y satisfying Eqs. (1) and ( 2 ) . Eqs. (1)and ( 2 ) can be rewritten by matrices. Let matrix
A be
(g :),
matrix X be
(i),
and matrix C be
(i:).
Eqs. (1) and
(2) can then be denoted as A X = C. By multiplying the inverse matrix of A , which is denoted as A - l , by both sides of A X = C from the left, we have A-IAX = A-lC, and then we have X = A-lC. Thus, if there is A - l , then the correct answer X =
(i)
is obtained. The existence of
A-l is guaranteed, since matrix A is regular. Gauss-Jordan's method [8] is usually used to compute A-l . According to this method, a given matrix is transformed into a unit matrix E =
(;:).
There are three classes of elementary transformations for matrices:
Ei(c) : multiply all elemeds in the ith column by c. Ei,j(c): add the ith column multiplied by c to the j t h column (i Pi,j : exchange the ith and j t h columns (i # j).
# j).
Since each elementary transformation is implemented by a matrix, the application of it is denoted as a multiplication of matrices. For example, if t is an elementary transformation and M is a matrix, then the application o f t to M is denoted as t M or M t . According to Gauss-Jordan's method, we have three concrete elementary transformations, t l , t2 and t ~where , tl = E1,2(-2), t 2 = E ~ , I ( - $ ) and t3 = Ez( By applying them to matrix A sequentially, we have t3 . t2 . tl . A = E. This implies that A-' = t3 . t 2 . tl.
i).
225
By multiplying A-l by both sides of A X = C from the left, we have
A-lAX = E . X = X =
= A-lC = t3 . t2 .ti .
Therefore, the correct answer to the given problem is 23 cranes and 12 tortoises. In the case of the simultaneous equations shown above, the problem consists of two matrices, A and C. Let the description d be a pair ( A ,C), and the answer a be a matrix X . The description and the answer are bound by a map, f : M2>2x M2*l + M2?l,that computes the value of X satisfying A X = C from ( A ,C), where Mi" is a set of all matrices that have i columns and j rows. Thus, it holds that a = f(d) and the answer a is correct. The elementary transformation tl transforms d into d2 = (tl . A , tl . C), t z transforms d2 into d3 = ( t z . tl . A , t z . tl . C ) , and at last t3 transformsda intodq = d'= ( t 3 ' t z . t l . A , t 3 . t Z . t l . C )= (E,A-'C). It is guaranteed that f(d) = f(d2) = ... = f(d'). Thus, the algorithm consists of a sequence of equivalent transformations. 2.2. Distributing Apples
Let us consider another problem. Suppose that two apples remain if apples are distributed to people in lots of three, that three apples remain if they are distributed in lots of five and that two apples remain if they are distributed in lots of seven. The number of people may vary in each distribution. What is the total number of apples? This problem can be expressed by the following congruence expressions: z z 2
(mod3)
(3)
zs3
(mod5)
(4)
z r 2
(mod7)
(5)
The symbol z is the total number of apples. There are in fact many numbers that satisfy Eqs. (3), (4) and ( 5 ) simultaneously. Such numbers are congruent with a modulus. The answer to the problem, then, is shown as a congruence expression, z k (mod m). This means that the correct answer, i.e., the total number of apples, is k, k m, k 2m, and so on. Congruence expressions are solved by using the Chinese Remainder Theorem [lo]. This theorem requires that all moduli are relatively prime. The moduli in Eqs. (3), (4)and (5), i.e., 3, 5 and 7, satisfy this requirement. By the definition of congruence expressions, Eq. (3) is rewritten as
+
+
226
x =2
+ 3t,
(6)
where t is an integer number. By substituting this z for Eq. (4), we have 2 3t G 3 (mod 5), and then we have 3t E 3 - 2 = 1 (mod 5). The correct answer to this is t E 2 (mod 5). This is obtained by substituting each element in the system of residues with modulus 5, ie., 0, 1, 2, 3 and 4 for t respectively. It holds that t = 2 5s, where s is an integer number. By substituting this t for that in Eq. ( 6 ) ,we have z = 2 + 3 x (2+5s) = 8+15s. This is rewritten as
+
+
2
E8
(mod 15 = 3 x 5).
(7)
Thus, by the series of procedures shown above, Eqs. (3) and (4) are incorporated into one congruence expression, Eq. (7). By applying a series of similar procedures to two expressions, Eqs. (5) and (7), they are incorporated into one congruence expression. Eq. (7) is rewritten as z =8
+ 15t’,
(8)
=
where t‘ is an integer. By substituting this z for Eq. (5), we have 8+15t’ 2 (mod 7), and then we have 15t’ 1 (mod 7). The correct answer to this is t’ E 1 (mod 7), and this is rewritten as t‘ = 1 + 7s‘, where s’ is an integer. By substituting this t’ for that in Eq. (8), we have x = 8 + 15 x (1 + 7s‘) = 23 105s‘. This is rewritten as
+
z E 23
(mod 105 = 3 x 5 x 7),
(9)
which is the correct answer to the problem. Therefore, the total number of apples is 23, 128 = 23 105, 233 = 23 2 x 105, and so on. In the case of the congruence expressions shown above, the problem consists of three congruence expressions, Eqs. (3), (4) and (5). Let the description d be a set {z E 2 (mod 3), z E 3 (mod 5), z 2 (mod 7)}, and the answer a be a congruence expression, x f k (mod m). The description and the answer are bound by a map, f : 2O + D, that computes values of k and m that satisfy all given congruence expressions in d, where D is a set of all congruence expressions. The original description d is transformed into d2 = {x G 8 (mod 15),x 2 (mod 7)}, and at last d2 is transformed into d3 = d’ = {z 23 (mod 105)). It is guaranteed that a = f ( d ) = f (d2) = f (d’). Thus, it holds that a = f (d) and the answer a is correct.
+
+
=
227
2.3. Fallible Greek This example is quoted from a book 171. The notations and terms used in this paper are the same as those used in the book [5]. There are four assumptions, A l , A2, A3 and A4, which are denoted by logical formulae.
A1
:
A2 :
A3 : A4 :
human(Turing) human(S0crates) greek(Socrates) ‘x.(human(x) 3 f a l l i b l e ( x ) )
Logical formulae A1 and A2 show that both Turing and Socrates are humans, and the logical formula A3 shows that Socrates is a Greek. The logical formula A4 shows that for any x if x is a human, then x is fallible. stands for an implication. For any two logical formulae X The symbol ‘2’ and Y , the notation X 3 Y is equivalent to the notation 1X V Y . The problem to be solved is to prove that there is a fallible Greek under these four assumptions. That is denoted by the following logical formula C.
c:
3u.(greek(u)A f a l l i b l e ( u ) )
After all, a logical formula to be proven is expressed by P as follows.
P:
(A~AA~AA~AAA~)>C
In general, a logical formula is proven by the resolution principle. The resolution principle can automatically prove logical formulae. In order to use the resolution principle, a logical formula to be proven must be described in a ‘clause set.’ In order to make a clause set from a logical formula, the logical formula must be in the Skolem canonical form, which is defined in Definition 2.1.
Definition 2.1. A logical formula in a form, 3x1 . . . 3z,.((F1 A . . . A F,) V
..
V
(GIA .. A G t ) ) ,
is in the Skolem canonical form, when each of F1,. . . ,F,, G1, . . . ,Gt is an atomic formula or a negation of an atomic formula, there is no variable other than xi,. . . ,xn and xi # x ~ for j i # j.
0
228
There is a general method for transforming a given logical formula into the Skolem canonical form [4].The original logical formula P is rewritten in the Skolem canonical form Q as follows.
Q:
3u.3z.(lhuman(Turing)V lhuman(Socrates) V lgreek(Socrates)
V (human(%)A -f a l l i b l e ( z ) ) V (greek(u) A f allible(u)))
For proving a logical formula, the resolution principle can be applied to a set of an instance of it, which is defined in 2.2.
Definition 2.2. Suppose that R is a logical formula in Skolem canonical form, then R is in a form 3 ~ 1 . - . . 3 ~ m .V(*A*1* V A , ) ,
where each Ai = Ai,l A . . .A A ~ , ,for ~ i = 1 , . . . ,n,and each Ai,j is an atomic formula or negation of atomic formula for j = 1,. . .,ni. By substituting a constant term for each variable z k (Ic = 1,.. .,m) in each A i j , an ‘instance’ A‘i,j is obtained. Let each A’i be a set {A’Q,. . . ,A‘i,,i}. A set {A’l, . . . ,A’,} is called a clause set of an instance of the logical formula R. 0
The resolution principle is defined in Definition 2.3.
Definition 2.3. Let S be a clause set of an instance of a logical formula. . . ,B,} as its Suppose that S includes C1 = {AI,. . . ,A,} and C2 = {BI,. elements. If there is an atomic formula A satisfying A E C1 and 1 A E C2, then a set (C, - { A } ) U (C2 - {lA}) is called a resolvent of S, and a set S U {(C, - { A } ) u (C2 - { - A } ) } is called a resolution of S . 0 Herbrand’s theorem [5] guarantees that if there is an instance of a logical formula in the Skolem canonical form and a clause set of the instance of the logical formula is proven to be true, then the logical formula is proven to be true.
229
The resolution principle guarantees that the truth value of a clause set of an instance of a logical formula and that of a resolution are the same. An empty set {} as an element of a clause set means a truth value ‘true.’ A logical formula A1 V . . . V A,V true is obviously true. Thus, a clause set R = {A,, . . . ,A,} is proven to be true, if there is a sequence R = R1, R2,. . . ,R,, where each Ri+l is the resolution of Ri for i = 1 , . . . ,m - 1 and R, has an empty set {} as its element. Again, let us consider the logical formula Q. In order to prove that Q is true by the resolution principle, it is necessary to obtain a clause set of an instance of Q. In order to obtain an instance of Q, each of variables u and z must be substituted for a constant term. The possible substitutions are [Turinglu] or [Socrates/u] and [Turing/z] or [Socrates/z]. Thus, there are four possible substitutions. A clause set of an instance Q[Socrates/z, Turing/u], for example, can not be proven to be true. For a substitution O = [Socrates/z, Socrates/u], an instance of Q is
Qe :
lhuman(Turing) V Thuman(Socrates) V Tgreek(Socrat es)
V (human(Socrates) A Tf a l l i b l e ( S o c r a t e s ) ) V (greek(Socrates) A f a l l i b l e ( S o c r a t e s ) ) .
In order to apply the resolution principle to QO, it is necessary to make a clause set of it. Let R1 be the clause set of the instance Qe. It is denoted by
R1 = {{lhuman(Turing)}, {ihuman(Socrates)}, { i g r e e k (Soc r a t e s)} , {human(Socrates),i f a l l i b l e ( S o c r a t e s ) } , {greek(Socrates), f allible(Socrates)}}.
If this R1 is proven to be true, then the original logical formula Q is also proven to be true. In order to prove that R1 is true, let us make a sequence R1, . . . , R, of resolutions and show that R, has an empty set {}. As the first step, the elements i g r e e k ( S o c r a t e s ) and greek(Socrates) are removed and then the resolution of R1 becomes the following R2. R2 = R1 U {{} U { f a l l i b l e ( S o c r a t e s ) } ) = R1 U {{fallible(Socrates)}}
230
As the second step, the elements T f allible(Socrates) and fallible(S0crates) are removed and then we have the resolution R3 of R2. R3
= R2 U {{human(Socrates)} U {}} = R2 U {{human(Socrates)}}
As the last step, the elements -human(Socrates) and human(S0crates) are removed and then the resolution of R3 becomes the following R4. R4
= R3 = R3
u {{I u 0 ) u (01
Since this R4 has the empty set {} as its element, the clause set R4 is true. Therefore, the clause set R1 of the instance &I3 is proven to be true, and then the original logical formula Q is also proven to be true. There is a fallible Greek. The name of the person is u = Socrates. 3. Discussion
The first and second examples are algebraic problems, and the third example is a logical problem. It has been shown that there is a common structure of algorithms for solving algebraic problems [6]. In this section, it is shown that there is also a common structure of algorithms for solving logical problems. 3.1. Common Structure of Algebraic Problems
For solving an algebraic problem, the problem is described by algebraic formulae, such as simultaneous equations and congruence expressions. Let the description be d. The description of the problem is bound to a correct answer, a. Let the map f be the relationship between d and a. It then holds that a = f(d). In general, it is difficult to compute f (d) directly. In both of the algebraic problems, the original description of the problem d is transformed into another form, d'. In the example of cranes and tortoises d is ( A , C ) and d' is (E, A-' . C). In the example of distribution of apples d is {x z 2 (mod 3 ) , x E 3 (mod 5),z 3 2 (mod 7)) and d' is {x E 23 (mod 105)). In both examples, it is easier to compute f(d') than t o compute f(d). In order to obtain a correct answer, it must hold that a = f(d) = f (d'). In
231
the case of simultaneous equations, the Gauss-Jordan’s method guarantees it. In the case of congruence expressions, the Chinese Remainder Theorem guarantees it. In conclusion, suppose that a description of a given problem is d and that a correct answer to the problem is a , then there is a map, f , and it holds that a = f ( d ) . The original description d is transformed into another description, d’. The transformation from d to d‘ may result in the formation of a sequence of transformed descriptions, d = dl , d2, . . . ,d, = d‘. A theorem guarantees that the equation a = f ( d ) = ... = f (d‘) holds. Thus, we obtain the correct answer. The algorithm for solving a given problem is a sequence of transformations that satisfy the equation. The equation and the transformations form a common structure of algebraic algorithms.
3.2. Common Structure of Logical Problems The algorithm used for solving a logical problem has a similar structure. In the example of fallible Greek, the problem is expressed by a logical formula, Q. Herbrand’s Theorem guarantees that Q is true if there is a substitution CJ and the instance Qu is true. There are four possible substitutions, 01,82,03,84. The original problem ‘whether the logical formula Q is true or not?’ is then changed to a new problem, ‘whether one of the instances & e l , Q02, Q03 and Q04 is true or not?’ The new problem is expressed by the set {Q81,Q02,Q03,Q04}. One of 01,02,03 and 04 is a substitution, 8 = [Socrates/z, Socrates/u], and later we see that Qe is true. We apply the resolution principle for solving the problem. The resolution principle requires that a target logical formula is expressed by a clause set. In the case of simultaneous equations, the problem is first expressed by simultaneous equations and later expressed by matrices. The notations used are different, but contents expressed are equivalent. Similar to the case of simultaneous equations, let all instances be expressed by clause sets in the case of logical formulae. Let IlQeiII be a clause set of Q0i for each i = 1,2,3,4. The description d of the problem to be solved is the set { ~ ~ Q 0IlQ0211, 1 ~ ~ ,11Q0311,llQ0,II}. The answer a can be one of true and false. The description and the answer are bound by a map, f : 2’ + {true,false}, that computes whether one of its elements is true or not. It holds that a = f ( d ) and the answer a is correct, where S is a set of all clause sets. According to the resolution principle, the truth value of a clause set of an instance and that of a resolvent of the clause set are equivalent. Thus, if
232
the element R1 of the description d is transformed into Ra, then we obtain a new description, d z , but the answer a is the same, i e . , a = f ( d ) = f ( d 2 ) . By applying the resolution principle sequentially, we have d3 and d4. It also holds that a = f ( d ) = f ( d 2 ) = f ( d 3 ) = f ( d 4 ) . Since the set R4 E d4 has an empty set as its element, it is obvious that R4 is true. Thus, we find f ( d 4 ) is true, and then obtain the answer a = f ( d ) = f ( d 4 ) = true. This is the correct answer. Thus, the structure of a logical algorithm is similar to the structure of algebraic algorithms.. There is a map binding a description of a given problem and a correct answer to the problem. The algorithm is a sequence of equivalent transformations. The description of a problem is transformed into another form where it is easy to compute the answer. Since the transformation is an equivalent transformation, the correct answer does not change. 4. Conclusion
Three examples of problem-solving are shown. Two problems are algebraic and one problem is logical. There is common structure in all of the examples. The equation a = f(d) always holds and an algorithm always consists of equivalent transformations. Algebraic method for solving are efficient, but a logical method is not. The existence of a common structure of algorithms suggests that there is an effcient method for solving for a logical problem.
Acknowledgment This study was supported by a grant from Tenshi College. The author is grateful to the anonymous reviewer for constructive suggestions, which helped t o improve the clarity of this paper.
References 1. Kiyoshi Akama, Tomokuni Shimizu, and Eiichi Miyamoto. Solving problems by equivalent transformations of declarative programs. Journal of Jsai., 13(6):944 - 952, November 1998 (in Japanese). 2. Edsger W. Dijkstra et al. Structured Programming. Science Sha, Tokyo, 1975 (translated into Japanese). 3. Kazuhiro Fuchi et al. Program Henkan. Kyoritsu Shuppan, Tokyo, 1987 (in
Japanese). 4. Masami Hagiya. Software Kagakv notameno Ronrigaku. Number 11 in The Iwanami Software Science Series. Iwanamishoten, Tokyo, January 1994.
233 5. Susumu Hayashi. Suri Ronrzgaku. Number 3 in Computer Sugaku Series. Corona Sha, Tokyo, 1989 (in Japanese). 6. Yuuichi Kawaguchi. An equivalent transformation paradigm that guarantees the quality of algorithms. In M. H. Hamza, editor, Proceedings of the ISATED International Conference: Artzficzal Intelligence and Soft Computing (ASC2001), pages 49 - 52, Cuncun, Mexico, May 2001. IASTED, ACTA Press. 7. Robert A. Kowalski. Logic for Problem Solving. Elsevier North Holland, Inc., 1979 (translated into Japanese). 8. Takeshi Onodera, Osamu Nakada, and Toshio Hashimoto. Kisokatei Senkei Daisugaku. Kyoritsu Shuppan, Tokyo, 1980 (in Japanese). 9. Joseph R. Shoenfield. Mathematical Logic. A K Peters, Ltd., Natick, Massachusetts, 1967. 10. Teiji Takagi. Shoto Seisuron Kogi. Kyoritsu Shuppan, Tokyo, second edition, 1971 (in Japanese).
234
GAMES ON GRAPHS: AUTOMATA, STRUCTURE, AND COMPLEXITY
BAKHADYR KHOUSSAINOV Computer Science Department T h e University of Auckland New Zealand email: bmk(0cs. oucklond.oc.nz
TOMASZ KOWALSKI* Japan Advanced Institute of Science and Technology Japan email: kowalski(0joist.ac. j p
1. Introduction and Basic Concepts Motivated by the work of Gurevich and Harrington [5], McNaughton [8] introduced a class of games played on finite graphs. In [8], McNaughton shows that all winnings strategies in his games can be implemented by finite state automata. McNaughton games have attracted attention of many experts in the area, partly because the games have close relationship with automata theory, the study of reactive systems, and logic (see, for instance, [13] and [12]). McNaughton games can also be used to develop gametheoretical approach for many important concepts in computer science such as models for concurrency, communication networks, and update networks, and provide natural examples of computational problems. For example, Nerode, Remmel and Yakhnis in a series of papers (e.g., 191, [lo]) developed foundations of concurrent programming in which finite state strategies of McNaughton games are identified with distributed concurrent programs. McNaughton games are natural descriptions of reactive systems in which the interaction between Controller (often referred to as Survivor) and Environment (Adversary) are modelled as certain two-player games. Winning *On leave from Department of Logic, Jagiellonian University, Poland.
235
conditions in these games can be thought of as specification requirements that Controller must satisfy. Winning strategies for Controller are thus identified with programs satisfying the specifications. Deciding whether or not Controller wins a given game can be seen as answering the question whether or not a given specification is realizable. If it is, then constructing a winning strategy amounts to synthesizing a correct controller program. Further, minimalization of the memory size of the winning strategy for Controller corresponds to the optimization problem of a correct controller. Again, we refer the reader to [12] for more details. Suppose you come across a McNaughton game. You will probably expect that the particular structure of the underlying system and the specification of winning conditions influence in some way the running times of algorithms that decide the game. Such an expectation is natural since many algorithms for deciding McNaughton games are not efficient and do not explicitly exploit either the structure of the underlying graphs or the form of winning conditions. An exception can be found in the paper of Zielonka [14] where it is shown that the winners of the McNaughton games have finite state strategies that depend on nodes that are called useful. The main purpose of this paper is to pursue this line of investigation a little further in a number of cases. In particular, we provide examples of classes of games for which the algorithms that decide these games explicitly use the nodes at which one of the players has more than one choice to make a move. We begin with the following definition extracted from [8],
Definition 1.1. A game r, played between two players called Survivor and Adversary, is a tuple ( S U A, E, R), where: (1) The sets S and A are disjoint and finite, with S being the set of positions for Survivor and A the set of positions for Adversary, (2) Theset Eofedgesissuchthat E g (Ax S)U(S x A )a n d fo ra lls E S and a E A there are a' E A and s' E S for which (s, a'), ( a ,s') E E, (3) The set R of winning conditions is a subset of 2SUA.
The graph G = (V,E), with V = S U A, is the system or the graph of the game, the pair R is the specification, and each set U E R is a winning set. In game r, a play (from PO) is an infinite sequence IT = p o , p l , . . . such that (pi,pi+l) E El i E w . Survivor always moves from positions in S , while Adversary from A. Define I n f (n)= { p I 3"i : p = p i } . Survivor wins the play T if Inf ( n ) E R; otherwise, Adversary wins n. We will refer
236
to finite initial segments of plays as histories. A strategy for a player is a rule that specifies the next move given a history of the play. Let f be a strategy for the player and p be a position. Consider all the plays from p which are played when the player follows the strategy f . We call these plays consistent with f f r o m p . We note that the definition above is not the original definition of McNaughton games. That is, in the original definition, one allows to pick a subset X of V , and fl is a collection of subsets of X . Then, in a play IT = p o , p l , . . . those nodes not in X are not counted toward satisfying a winning condition. This distinction is important since Nerode, Remmel, and Yakhnis in [9] prove that McNaughton games can be solved in time proportional to 21xl . . IEllXl!). In other words, when X is very small compared t o V , all McNaighton games can be solved efficiently.
(1x1
Definition 1.2. The strategy f for a player is a winning strategy from p if the player wins all plays from p consistent with f . In this case the player wins the game from p. To decide game r means to find the set of all positions, denoted by Win(S),from which Survivor wins. The set Win(A) is defined similarlya. jF'rom the definitions above it is clear that all graphs we consider are bipartite and directed. It is customary in the context of games to refer to members of V as nodes rather than as more graph-theoretical vertices. For a node w of a graph B = (V,E), we write Out(w) = { b 1 ( v , b ) E E } and In(.) = { b 1 (b, w) E E } . Usually nodes of set A are denoted by a , and of set S by s, possibly with indices. As we have already said, McNaughton's algorithm in [8] that decides games is inefficient. In [9] Nerode, Remmel and Yakhnis improved the algorithm by deciding any given game F in O(IV(!21vlIVllEl)-time which is, of course, still far from being efficient. S. Dziembowski, M. Jurdzinski, and I. Walukiewicz in [2] investigated questions related to the size of memory needed for winning strategies. In particular, they prove that for each n there is a game r such that the size of V is O(n) and the memory size for finite state winning strategies for these games are at least n factorial. A related question is t o find conditions on either the specifications or the system which ensure that the games can be decided efficiently and the memory size for winning finite strategies are sufficiently small. While the present ~~
aAny McNGghton game I? is a Bore1 game. Hence, by the known result of Martin (see [7]), r is determined. Therefore W i n ( S )U W i n ( A )= S U A .
237
paper has some bearing on the above question, it is also a continuation of a research trend, which we briefly summarise in the next paragraph. Dinneen and Khoussainov have used McNaughton games for modelling and studying structural and complexity-theoretical properties of update networks (see [l]).A game r is an update game if R = { V } . The system (V,E ) is an update network if Survivor wins the update game. Speaking informally, Survivor is required to update [i.e., visit) every node of the system as many times as needed. In [l]it is shown that update games can be decided in O(lVl(lVl+IEl))-time. Update games have been generalized in [3] t o games in which the specification R contains more than one set. Namely, a game I' is a relaxed update game if U n W = 0 for all distinct U , W E R. It is proved that there exists an algorithm that decides relaxed update games in O(IV12(IVI IEl)))-time. In [6]Ishihara and Khoussainov study linear games in which R forms a linear order with respect to the set-theoretic inclusion. They prove that linear games can also be decided in polynomial time with parameter IRI.
+
Clearly, in the results above, all the constraints are specification constraints. In other words, the games are described in terms of certain properties of specifications from R. In addition, the results as they stand-and most of their proofs-do not explicitly show the interplay between the structure of the underlying systems (V,E ) , the running times of algorithms that decide games, and the specifications in R. We try to bridge this gap by explicitly showing how running times of algorithms that decide certain classes of games depend upon the structure of the systems and specifications. Here is a brief outline of the paper. In the next section, we introduce no-choice games and present a simple algorithm that decides them in time linear on the size of the game. We also provide a result that shows how the structure of the no-choice games is involved in finding finite state winning strategies with a given memory size. In Section 3 we revisit update games, and provide a n algorithm that explicitly uses information about the number of nodes at which Adversary can make a choice, i.e., the members of A with at least 2 outgoing edges. In Section 4, we consider games in which specifications in 52 are closed under union. For such union-closed games we provide a decision algorithm whose running time depends explicitly on some structural information about the underlying systems of games. We note that the main result of this section can be obtained from the determinacy result of Zielonka [14]. However, our proof is direct and simple and does not need to employ the full strength of Zielonka's determinacy theorem.
238
The final section discusses some issues for future work. 2. No-choice games
We start off with games where the structure of the system forces one of the players to always make a unique choice at any given node of the player. Without lost of generality we can assume that this player is Adversary. Formally :
Definition 2.1. A no-choice game is a McNaughton game = (V,E , R) such that for all a E A, s1,sg E S if ( a , s l ) ,( a , s 2 ) E E then s1 = s2. No-choice games are one player games, with Survivor as the sole player, because Adversary has no effect on the outcome of any play. Below we provide a simple procedure deciding no-choice games by using Tarjan's algorithm that detects strongly connected directed graphsb. The algorithm is simple but shows a significant difference between the time needed to decide McNaughton games in the general case and in the case of no-choice games. The following is a simple observation.
Lemma 2.1. I f X is a strongly connected component of G in a no-choice game r = (V,E , R) and > 1, then O u t ( a ) c X for every a E A n X .
1x1
0
Let r = (S U A, E , R) be a no-choice game. Call a winning set U E R S-closed, if O u t ( a ) & U for every a E A n U , and Out(s)n U # 0 for every s E S n U . Clearly, if 7r is a play won by Survivor in game r then Inf(7r) must be S-closed. Thus, the following lemma holds true:
Lemma 2.2. Survivor wins the no-choice game I? if and only if Survivor wins the game r' which arises from 'I by removing all not S-closed winning 0 sets. Let U E R be an S-closed winning set. Consider the game r ( U ) whose graph is the restriction to U of the graph of I?, and whose set of winning conditions R(U) is { U } . Define the graph G ( U ) = ( V ( U ) , E ( U ) ) , where V ( U ) = S n U , and (z,y) E E ( U ) if and only if z,y E V ( U ) and (z, a) , (a, y) E E for some a E U n A. Thus, in graph G(U) there is a path bA graph Q = (V,E) is strongly connected if there is path between any two nodes of the graph. Tarjan's algorithm detects whether or not the graph Q is strongly connected in O(lVl IEl)-time
+
239
between nodes p and q if and only if there is a finite play s1, al, . . . , a,-1, s, in r ( U ) such that p = s1 and q = sn. The following is easy: Lemma 2.3. Survivor wins r ( U ) ifl the graph G ( U ) is strongly connected. 0
Now we are ready to prove the following theorem: Theorem 2.1. There exists an algorithm that decides any given no-choice game I' = (V, E , R) in O(lR] . (IVl JE]))-time.
+
Proof. Let p be a node in V. Here is a description of a desired algorithm:
If there is no S-closed U E R then declare that Survivor loses. If none of these graphs G ( U ) for S-closed U E R is strongly connected, then declare that Survivor loses. Let X be the union of all S-closed U E R such that the graph G ( U ) is strongly connected. Check whether or not there is a path from p into X . If there is no path from p into X then declare that Survivor loses. Otherwise, declare that Survivor wins. It takes linear time to perform the first part of the algorithm. For the second part, use Tarjan's algorithm for detecting strongly connected graphs. Namely, for each S-closed set U apply Tarjan's algorithm to check if G ( U )is strongly connected. Hence the overall running time for the second part is proportional to \R(. (IV( ( E l ) .For the third part, constructing X and checking if there is a path from p to X takes linear time. Thus, the algorithm runs at most in O(lRl . ([El+ IVI)-time. The correctness of the algorithm is clear. 0
+
Thus, the proof of Theorem 2.1 shows that deciding no-choice games is essentially dependent on checking whether or not the graphs G ( U ) = (V(U)E , ( U ) ) ,where U is S-closed, are strongly connected. Therefore we single out the games that correspond to winning a single set U E R in our next definition: Definition 2.2. A basic game r consists of a directed graph 4 and player
Survivor, where Survivor is the sole player. Given a basic game r, a play from a given node vo is a sequence T = vo, ~ 1 , 7 1 2 , .. . such that (vi,vi+l) E E for all i E w. Survivor wins the play ) V. Otherwise, Survivor looses the play. Thus, Survivor wins if I n f ( ~= the basic game I? iff the graph G is strongly connected.
240
Let r be a basic game. Our goal is to find finite state strategies that allow Survivor to win the game. For this, we need to formally define finite state strategies. Consider an automaton A = (Q, qo, A, F ) , where V is the input alphabet, Q is the finite set of states, qo is the initial state, A maps Q x V to Q, and F maps Q x V into V such that ( v , F ( q , v ) )E E for all q E Q and v E V . The automaton A induces the following strategy, called a finite state strategy. Given v E V and s E Q, the strategy specifies Survivor’s next move which is F ( s ,v). Thus, given uo E V , the strategy determines the run r(vo,A) = wo,v1,v2,..., where vi = F(qi-1,wi-l)and qi = A(vi-1,qi-l) for each i > 0. If Inf(r(v0,A)) = V , then A induces a winning strategy from WO. When Survivor follows the finite state strategy induced by Adversary, we say that A dictates the moves of Survivor. To specify the number of states of A we give the following definition.
Definition 2.3. A finite state strategy is an n-state strategy if it is induced by an n state automaton. We call 1-state strategies no-memory strategies. The next result shows that finding efficient winning strategies in basic games is computationally hard. By efficient winning strategy we mean an n-state winning strategy for which n is small.
Proposition 2.1. For any basic game F, Survivor has a no-memory winning strategy if and only if the graph G = (V,E ) has a Hamiltonian cycle. Therefore, finding whether or not Survivor has a no-memory winning strategy is NP-complete. Proof. Assume that the graph G has a Hamiltonian cycle V O ,. . . ,v,. Then the mapping wi + vi+.l(mo~(n+l))establishes a no-memory winning strategy for Survivor. Assume now that in game I? Survivor has a nomemory winning strategy f . Consider the play r = po, pl,pp, . . . consistent with f . Thus f(pi) = pi+l for all i. Since f is a no-memory winning strategy we have I n f ( r ) = V . Let m be the least number for which po = p,. Then V = {PO,. . . , p , } as otherwise f would not be a winning strategy, and 0 hence the sequence PO,.. . , p , is a Hamiltonian cycle.
It is not hard to see that there exists an algorithm running in 0(lVl2)time that for any given basic game in which Survivor is the winner provides an automaton with at most IVl states that induces a winning strategy (just check if for all 2,y E V there are paths connecting x to y and construct
241
a desired automaton) . Therefore, the following seem natural: If a player wins the game a t all, then how much memory is needed to win the game? For a given number n, what does the underlying graph look like if the player has a winning strategy of memory size n? We will provide answers, however, we will not give full proofs. Full proofs can be found in [4]. Our goal is t o analyze the case when n = 2, that is when Survivor has a 2-state winning strategy, as the case for n > 2 can then be derived without much difficulty. The case when n = 1 is described in Proposition 2.1. The case when n = 2 involves some nontrivial reasoning provided below. The case when n > 2 can be generalized easily (see [4]).
Case n = 2. We are interested in graphs B = (V,E ) such that IIn(q)l 5 2 and IOut(q)I 5 2 for all q E V . A path p l , . . . , p , in graph B is called a 2-state path if lIn(p1)I = IOut(p,)] = 2 and IIn(pi)l = ] O u t ( p i ) J= 1 for all i = 2,. . . ,n - 1. If a node q belongs to a 2-state path then we say that q is a 2-state node. A node p is a 1-state node if IIn(p)I = IOut(p)I = 1 and the node is not a 2-state node. A path is a 1-state path if each node in it is a 1-state node and no node in it is repeated. We now define the operation which we call the Glue operation which can be applied t o a finite graph and a cycle to produce another finite graph. By a cycle we mean any graph isomorphic to ( { c l ,. . . ,cn}, E ) , where n > 1 and E = {(c~,cz),. . . , ( ~ ~ - c,), 1 , (c,, Q)}. Assume that we are given a graph 6 = (V,E ) and a cycle C = (C,E ( C ) ) so that C n V = 0. Let P I , . . . , P, and Pi, . . . , PA be paths in 4 and C, respectively, that satisfy the following conditions: 1) The paths are pairwise disjoint; 2) Each path Pi is a 1-state path; 3) For each i = 1 , . . . ,n, we have lPil = lP,!l.The operation Glue has parameters B, C, PI,. . . , P,, Pi, . . . , PA defined above. Given these parameters the operation produces the graph E'(V',E') in which the paths Pi and Pi are identified and the edges E and E ( C ) are preserved. Thus, one can think of the resulting graph as one obtained from B and C so that the paths Pi and Pi are glued by putting one onto the other. For example, say PI is the path pl,p2,p3, and Pi is the path p',,p&,p$. When we apply the operation Glue, PI and Pi are identified. This means that each of the nodes pi is identified with the node p:, and the edge relation is preserved. Thus, in the graph B' obtained we have the path { p l , p ; } , { p ~ , p $ }{ p, s , p $ } . It is easily checked that in the resulting graph 4' each of the paths Pi is now a 2-state path. Definition 2.4. A graph B = (V,E ) has a 2-state decomposition if there is a sequence ( G l , C l ) , . . . , ( B n , C n ) such that 61 is a cycle, each
is obtained from the Gi and Ci,and G is obtained from G, and C n by applying the operation Glue.
Gi+l
An example of a graph that admits a 2-state decomposition can be given by taking a union C1,. . . , C, of cycles so that the vertex set of each Ci,i = 1,.. . ,n - 1, has only one node in common with Ci+l and no nodes in common with other cycles in the list.
Definition 2.5. We say that the graph G = (V,E ) is an edge expansion of another graph 6’ = (V‘,E’) if V = V’ and E’ E. The following theorem provides a structural characterization of those strongly connected graphs which Survivor can win with 2-state winning strategies.
Theorem 2.2. 141 Survivor has a 2-state winning strategy in a basic game I’ = (G, { V } )i f and only if 6 is an edge expansion of a graph that admits a 2-state decomposition. 3. Update Games Revisited
Recall that a game of type r = (V,E , { V } )is called a n update game; and is an update network if Survivor wins the game. In this section all the games considered are update games. Our goal here is twofold. On the one hand, we describe a decomposition theorem for update networks. For a full proof of this theorem we refer the reader to [l].On the other hand, we provide a new algorithm for deciding update networks so that the algorithm runs in linear time on the size of the graph given a certain set of of Adversary nodes as a parameter. More formally, let r = (V,E , { V } )be an update game. Let C be the set of all Adversary’s nodes a such that IOut(a)l > 1. In other words, C contains all nodes at which Adversary has a choice of at least two different moves. We provide an algorithm deciding update games, so devised that its running time shows what role the cardinality of C plays in the decision procedure. Namely, our algorithm depends on the parameter ICI and runs in the time I C . ( I V l f IEI), where Ic depends on ICI linearly. Let r = ( V , E , { V } )be an update game. For any s E S define Forced(s) = { a I (Out(a)(= l&(a,s) E E } . Thus, Forced(s) is the set where Adversary is ‘forced’ to move to s. Note the following two facts. If r = ( V , E , { V } )is an update network then for every s E S , the set Forced(s) is not empty. Moreover, if IS( 2 2, then for every s E S there exists an s’ # s and a E Forced(s) such that (s‘,a) E E. Next we give the
243
definition of forced cycle in a McNaughton game which will play a key role in our analysis of update games.
Definition 3.1. In a game r, a forced cycle is a cycle ( a k , s k , . . . ,a2, sg, a l , s1) such that ai E Forced(si) and si E s. Forced cycles have even length, and are fully controlled by Survivor. Using the facts above one now can show that any update network r with IS[ > 1 has a forced cycle of length 2 4. The lemma below tells us that forced cycles can be used to reduce the size of the underlying graph and obtain an equivalent game.
Lemma 3.1. Let I? be a n update game with a forced cycle C of length 2 4. W e can construct a game r' with IV'I < (VI such that r is a n update network iff I?' is one. Proof (sketch). We construct the graph ( V ' ,E') for I". Consider C = ( a k , s k , . . . ,a2, s g , a l ,s1). For new vertices s and a define S' = (S \ {SI, . . . ,s k } ) U {s} and A' = ( A \ { a l , . . . ,a k } ) U { a } . The set E' of edges consists of all the edges in E but not the edges in C , edges of the type ( s ,a ' ) if (si,a') E E , or (a',s) if ( a ' , s j ) E E , or (s',a) if ( s ' , a k ) E E for some si,s j , a k E C. We also put ( a ,s ) and (s,a ) into E'. Thus, the cycle C has been reduced. It is routine to show that r is an update network iff r' is one. The idea is that Survivor controls C fully. The operation of producing I" from I' and a forced cycle C is called the contraction operation. In this case we say that I? is an admissible extension of r'. Thus, for I" to have an addmisible extension I" must possess a forced cycle of length 2. Clearly, there are infinitely many admissible extensions of I".
Definition 3.2. An update game I' = (G, { V } )has a forced cycle decomposition if there exists a sequence rl, . . . , r n such that IS11 = 1, IOut(sl)l = ]All, where S1 = {sl}, and each ri+l is an admissible extension of ri,and rn = I'. The sequence rl, . . . , rn is called a witness for the decomposition. Using the lemma and the definition above one can prove the following theorem. The theorem gives us a complexity result one the one hand, and a description of update networks on the other (see [l]).
244
Theorem 3.1. There exists an algorithm that given a game I? decides in O(lVllEl) time whether or not the game is an update network. Moreover, an update game I? is an update network i f and only i f it has a forced cycle decomposition. We note that Nerode, Remmel, and Yakhnis in [9]prove a more general complexity theoretic result. They prove that if R is of the form {Y 1 21 Y E 2 2 } , where 21 and 2 2 are fixed subsets of V , then such McNaughton games (called interval games) can be solved in time proportional to (21211 1)JEJ.Clearly, update games are interval games. Thus, the complexity theoretic result in the theorem above is coverred by the result of Nerode, Remmel, and Yakhnis. The theorem above, however, provides a structural property of update networks, and tells one how update networks can be built.
+
Now we show how the set C = { a E A 1 IOut(a)I > 1) can be used to decide update games. Our algorithm shows that if the cardinality of C is fixed then update games can be decided in linear time. Let X be a subset of V in a game I? = (G, 0). The graph G X is defined as the subgraph of (2 whose vertex set is V \ X . We begin with the following simple lemma.
Lemma 3.2. Assume that C is a singleton and C = {a}. If Survivor wins I' then In(.) # 8 and Out(a) is contained in a strongly connected component of (2{,}. Proof. It is clear that In(.) # 8 as otherwise a would not be visited infinitely often in each play. Assume now that no strongly connected component of G{,} contains Out(a). There are z,y in Out(a) such that the graph 91,) does not contain a path from x into 9. Consider the strategy that dictates Adversary to always move to x from the node a. Then, for any play T consistent with this strategy, Inf ( T ) does not contain y . Hence, Survivor cannot win I?. We will generalize the lemma above to the case when the cardinality of C is greater than 1. In other words, Adversary has more than one node at which a choice can be made. Let a l , . . . ,a, be all the nodes from C.
Lemma 3.3. Assume that Survivor wins I?. Then the following two properties hold true: (1) Each set In(ai) is not empty for i = 1,.. . ,n.
245
(2) There is an element b E C such that Out(b) is contained in a strongly connected component of Gc. Proof. The first property is clearly true as otherwise Survivor could not win the update game r. We show how to prove the second property. Take a1 . Assume that no strongly connected component of Gc contains Out(a1). Then there are X I and y1 in Out(a1) such that GC does not contain a path from x1 to y 1 . We make the following claims:
Claim 1. There is an i > 1 such that for every z E Out(ai) there is a path from z t o y1 in the graph Gc. In order t o prove the claim assume that for each ai, i > 1, there is a zi such that there is no path from zi into y1 in the graph Gc. Define the following strategy for Adversary. Any time when a play comes to ai, i > 1, move to zi. At node a1 move to X I . It is not hard to see that in any play T consistent with this strategy the node y1 does not belong to Inf ( T ) . This contradicts the fact that Survivor wins r. The claim is proved. Without loss of generality we can assume that a2 satisfies the condition of the claim above. If Out(a2) is contained in a strongly connected component, then the lemma is proved. Otherwise, there are x2,y2 E Out(a2) such that the graph GC does not have a path from 5 2 to y 2 . We now prove the following.
Claim 2. There is an i with 1 5 i 5 n such that for every z E Out(ai) there is a path from z to y2 in the graph Gc. Moreover, for any such i it must be the case that i > 2. Assume that a1 satisfies the claim. Then in Gc ' there is a path from X I to y 2 . Since a2 satisfies Claim 1, in GC there is a path from y2 to y1. Then, the path from x1 through y2 to y1 is in Gc as well. This is excluded by our initial assumptions about al. Thus, i # 1. Certainly a2 cannot satisfy Claim 2 either. Then, we complete the proof of Claim 2 by repeating the argument we employed to prove Claim 1. Now, repeating inductively the above reasoning, and suitably renumbering nodes, we may assume that the sequence a l , . . . ,aj has the following properties: (1) In each Out(ak), k = 1,.. . , j - 1, there are xk,yk such that the graph Gc contains no path from xk to y k . (2) For all z E Out(ak) with k = 2 , . . . , j there is a path from z to y k - 1 in the graph g c .
246
Now, if the set Out(aj) is not contained in a strongly connected component of GC then there is an a E C such that for all z e Out(@)there is a path from z to yj. Otherwise, one can again show that Adversary wins the game by never moving to yj. Indeed, the assumptions above guarantee that all paths from z j to y j go through an Adversary's node. Therefore Adversary can avoid visiting the node yj. This, however, contradicts the assumption that Survivor wins the game. Moreover, as above, it can be shown that a 6 { a l , . . . ,u j } , It follows that j < n. Thus, we can conclude that there is an i 5 n such that Out(ai)is contained in a strongly connected 0 component . The lemma is proved. By virtue of the lemma above we can pick an a E C such that Out(a) is contained in a strongly connected component of E c ; denote the component by X,. We construct a new update game r' = (V',E', { V ' } )as follows: (1) V' = (V \ X,)U { s}, where s is a new Survivor's node. (2) E' = ( E n v r 2u) { ( s , u ) I 3t E x,((t,a) E E } u { ( a , ~ I) 3t E x a ( ( a , t )E E ) ) .
We refer t o r' as the reduced game. The following lemma shows the use of reduced games. Lemma 3.4. Survivor wins the game l' if and only if Survivor wins I?'.
Proof. Let f be Survivor's winning strategy in I?. We describe a winning strategy f ' for Survivor in I" which simulates f . When the play takes place outside {s}, then f ' mimics the moves dictated by f for nodes outside X,. When the play arrives at s then Survivor scans f forward up to the nearest point where f leaves X,. Obviously such a point exists. Suppose f does so by requiring a move to a node y 6X,. Then in the game r' Survivor also moves to y. It is not hard to see that f' thus described is indeed a winning strategy. Now assume that f ' is Survivor's winning strategy in r'. We describe Survivor's winning strategy f in I? by simulating f'. When the play takes place outside X, then f mimics f'. When the play arrives a t X,, the strategy f tells Survivor to: (1) visit each node of X,, then (2) find node y t o which Survivor moves in game cording to strategy f', then
I" from node
s ac-
247
(3) find z E X , such that (z,y) E E , and move to finally, (4) from z move to y.
2
inside X,, then,
It is clear that f thus described is well-defined, i.e., Survivor can do what f requires. That f is indeed a winning strategy is not hard to see either. 0
Assume that an a E C is such that Out(a) is contained in a strongly connected component Gc. Consider the reduced game I",and its underlying graph 6' = ( V ' , E ' ) . The natural mapping h : V + V' defined by putting h(v) = s for all w E X,, and h(w) = w otherwise, satisfies the following properties:
h is onto; for all z,y E V , (z, y) E E and z,y $ X , implies that ( h ( z ) h, ( y ) )E E' ; X is a strongly connected component of Gc if and only if h ( X ) is strongly connected component of G&. These observations together with Lemma 3.3 yield that if Survivor wins then there is an a E C such that Out(a) is contained in a strongly connected component of GL. Moreover, by Lemma 3.4, we can reduce the sizes of strongly connected components to singletons one by one always arriving at an equivalent game. This amounts to a proof of the following lemma.
r'
Lemma 3.5. If Survivor wins the update game r then for any a E C the set Out(a) is contained in a strongly connected component of Gc. 0 Now we are ready to prove a theorem.
Theorem 3.2. There exists an algorithm that, given an update game r with ICI = n, decides whether or not r is an update network in running time proportional to n . (IV(+ IEI).
Proof. Our procedure uses Tarjan's algorithm. We describe the basic steps of our procedure. Its correctness follows from previous lemmas. (1) If C = 0 then apply Tarjan's algorithm to see if G is strongly connected. If 6 is strongly connected then Survivor wins; otherwise Adversary wins.
248
(2) Find all strongly connected components, X I , . . . , Xm, of Qc by using Tarjan’s algorithm. (3) If for some a E C the set Out(a) is not contained in one of the strongly connected components X I , . . . , X m , then Adversary wins. (4) Construct the graph G(C)= ( V ( C ) , E ( C )as ) follows: (a) V‘ = (V \ UaEcXa)U ( ~ 1 , .. . ’sk}, where each si is a new Survivor’s node, and k = ICI. (b) E‘ = ( E n V”) U u;=,{(Si,a) I 3t E X a i ( ( t , a i ) E E } U U;=,{(a,si) I 3 E L ( ( a , t )E ( 5 ) If Survivor wins the no-choice game r(C) = (G(C),{ V ( C ) } )then , Survivor wins the game I?. Otherwise, Adversary is the winner.
a).
It is not hard t o see that the algorithm runs in the time required.
0
4. Union-Closed Games In this section, we focus on union-closed games, that is the games in which the specification set R is closed under the set-theoretic union operation. Structurally, it is a natural property, and we will use it in an essential way in the algorithm deciding these games. Let r be a union-closed game. Consider a E A such that IOut(a)I > 1. Let So and S1 be pairwise disjoint nonempty sets that partition Out(a>.We define two games ro and rl, where = V , Ri = R, and Ei = E \ { ( a ,s) I s E Si}. In other words, in game ri, moves of Adversary at node a are restricted to Si. Here is the main theorem from which we will deduce an algorithm for deciding union-closed McNaughton games.
Theorem 4.1. Let r be a union-closed game. Survivor wins the game I’ from p if and only if Survivor wins each of the games ro and rl from p . Proof. We need to prove the nontrivial direction. Let fo and f 1 be winning strategies of Survivor in games ro and rl,respectively. We construct the following strategy f for Survivor in the original game I’. Survivor that begins its play by first emulating fo. Assume that p , p l , . . . , p n , a is the history of the play so far, and Survivor is emulating the strategy f E , where E E {0,1}. Now consider Adversary’s move from a. There are two cases.
Case 1. Adversary moves into S,. In this case, Survivor emulates fE until a is reached again. Case 2. Adversary moves into Sl-, by choosing an s E Sl-,.In this case, Survivor scans the history h = p , p l , . . . ,p,,a,s and “projectd’ it
249
into the game rl-e. The “projection” is obtained from h by forgetting all the detours that belong to the game re.More formally, Survivor does the following:
0
0
scans h from the beginning up to the first appearance of a followed by a t E S,; keeps scanning h up to the next appearance of a (there must be such, because h ends with a followed by s @ Sf); forms h‘ by identifying the two appearances of a in the the sequence a, s, . . . ,a and cutting off everything in between; repeats the procedure until the end of h.
The “projection” h’ obtained this way will be a history of a play from
rl-€.The next move of Survivor then coincides with the move of Survivor in the game step after h’.
required by the the winning strategy
fl-€
for the next
This strategy is a winning strategy. Indeed, consider a play IT consistent with this strategy. If after a certain history h of 7r Adversary always moves to S,from a then the play IT‘, obtained from IT by removing the initial segment h, is a play in I?€. Then, Survivor wins IT by resorting to the strategy fe after h has been completed. By symmetry, Survivor also wins any play 7r in which Adversary almost always moves to S I - ~Assume . that Adversary switches infinitely often from SOto S1 and back during the play. Then IT can be written as IT = a l P l a z P 2 . . . , where TI = ala2.. . is a play in l?o consistent with f o and 7r2 = POPI.. . is a play in rl consistent with f 1 . Therefore, Inf(7r) = Inf(r1)U Inf(n2). Since fo and f 1 are winning strategies for Survivor, we must have Inf (TI),Inf ( ~ 2 E ) R. By unionclosedness, we get Inf ( I T ) € R. Thus, f is the winning strategy for Survivor as required. 0 As a corollary we obtain a complexity-theoretic result for deciding union-closed games. To formulate it, we need yet another definition.
Definition 4.1. Let I? = (V,E,R) be a game. An instance of r is any game I?‘ = (V‘,E’, a‘) such that V’ = V , 0‘ = s2, and E’ C E such that for every a E A the set Out(a) with respect to E‘ has cardinality 1. Now we can state: Theorem 4.2. Let I? = (V,E , s2) be a union closed game. Let a l , . . . , ak be all nodes in A such that ni = IOut(ai)) > 1, i = 1 , . . . ,k. Then the following is true:
250
(1) Survivor wins I? if and only if Survivor wins every instance of .'I (2) Deciding the game r takes O(nl . . . . n k . (01. (IVl -k IEl)-time. +
Proof. Part 1 follows from Theorem 3. Part 2 follows from Theorem 1, the first part of the theorem, and the fact that there are exactly nl . . . . .n k instances of r. 0 Corollary 4.1. If Survivor looses a union-closed game, then Adversary has a no-memory winning strategy. Proof. By the theorem above, Adversary wins an instance of the game. Such an instance is itself a no-choice game in which Adversary wins, and 0 the strategy naturally derived is a no memory strategy.
We note that the corollary above can be obtained from the known determinacy result of Zielonka [14]. However, our proof is direct and simple and does not need to employ the full strength of Zielonka's determinacy theorem. 5 . Concluding Remarks
In this paper we have shown that McNaughton games can be studied by exploiting the relationship between specifications and the structure of the underlying graphs. This seems to be a natural approach if one wants to find efficient algorithms for deciding different classes of McNaughton games and have practical implementations of winning finite state strategies. The ideas presented in this paper can clearly be generalized and produce new algorithms for deciding McNaughton games. For example, we plan to investigate the question how the cardinality of the set at which Adversary has more than one choice to make a move can affect the complexity of decision algorithms for McNaughton games.
References 1. M. J. Dinneen and B. Khoussainov. Update networks and their routing strategies. In Proceedings of the 26th International Workshop on Graph-Theoretic Concepts in Computer Science, WG2000, LNCS 1928, p 127-136, 2000. 2. S. Dziembowski, M. Jurdzinski, and I. Walukiewicz. How Much Memory Is Needed to Win Infinite Games? in Proceedings of Twelfth Annual Symposium on Logic in Computer Science (LICS 97), p.99-118, 1997. 3. H.L. Bodlaender, M.J. Dinnccn and B. Khoussainov. On Game-Theoretic Models of Networks, in Algorithms and Computation (ISAAC 2001 proceedings), LNCS 2223, P. Eades and T. Takaoka (Eds.), p. 550-561, Springer-Verlag Berlin Heidelberg 2001.
251
4. B. Khoussainov. Finite State Strategies in One Player McNaughton Games. Proceedings of the 10th International Conference of CDMTCS, France, to appear. 5. Y. Gurevich and L. Harrington. Trees, Automata, and Games, STOCS, 1982, pages 60-65. 6. H. Ishihara, B. Khoussainov. Complexity of Some Games Played on Finite Graphs. Proceedings of the 28th International Workshop in Graph-Theoretic Conspets in Computer Science, Ed. L. Kucera, LNCS 2573, p.270-282, 2002. 7. D. Martin. Bore1 Determinacy. Ann. Math. Vol 102, 363-375, 1975. 8. R. McNaughton. Infinite games played on finite graphs. Annals of Pure and Applied Logic, 65:149-184, 1993. 9. A. Nerode, J. Remmel, and A. Yakhnis. McNaughton games and extracting strategies for concurrent programs. Annals of Pure and Applied Logic, 78:203242, 1996. 10. A. Nerode, A. Yakhnis, V. Yakhnis. Distributed concurrent programs as strategies in games. Logical methods (Ithaca, NY, 1992), p. 624-653, Progr. Comput. Sci. Appl. Logic, 12, Birkhauser Boston, Boston, MA, 1993. 11. R.E. Tarjan. Depth first search and linear graph algorithms. SIAM J. Computing 1:2, p. 146-160, 1972. 12. W. Thomas. On the synthesis of strategies in infinite games. in: STACS 95 (E.W. Mayr, C. Puech, Eds.), Springer LNCS 900, 1-13, 1995. 13. M. Vardi. An automata-theoretic approach to linear temporal logic. Proceedings of the VIII Banff Higher Order Workshop. Springer Workshops in Computing Series, Banff, 1994. 14. W. Zielonka. Infinite games on finitely coloured graphs with applications to automata on infinite trees. Theoretical Computer Science, 200, 135-183, 1998.
252
COMPUTATIONAL COMPLEXITY OF FRACTALS
KER-I KO Department of Computer Science State University of N e w York at Stony Brook Stony Brook, N Y 11794 E-mail: [email protected] This paper surveys recent work on the computability and computational complexity of two-dimensional regions and fractal curves, in the context of Turing machinebased computable analysis and discrete computational complexity theory.
1. Introduction
Fractals are interesting objects that occur in many different areas of research (see, e.g., Mendelbrot [ll],Peitgen et al. [12], Barnsley [l]and Falconer [5]). Simple iterative algorithms can produce complicated fractals in the limit. These fractals, though easy to generate, often have fine structures that make their basic properties hard to compute. For instance, Mendelbrot Ill], in his study of the coastline problem, pointed out that it may be easy to generate a Koch-like two-dimensional fractal curve but hard to measure its length. In order to understand the precise relations between the computational complexity of a fractal curve itself and that of its other properties, such as its length and the area of its interior, we need to develop a formal theory of computational complexity of fractals. In this paper, we present a short survey of recent work toward this goal. We apply the formal theory of computable analysis and computational complexity of real functions, which use Turing machines as the basic computational model, to the study of fractals. In particular, we discuss how to extend these theories to define computable and feasibly computable twodimensional regions. Based on this computational model, we then consider some basic issues about fractals, including the following questions: (1) If a two-dimensional curve I? is computable (or polynomial-time computable), does it follow that its length is also computable (or, respectively, polynomial-time computable)?
253
(2) If a two-dimensional Jordan curve (i.e., a simple closed curve) I' is computable (or polynomial-time computable), does it follow that the area of its interior is also computable (or, respectively, polynomialtime computable)? (3) If a two-dimensional curve I? is computable (or polynomial-time computable), does it follow that its Hausdorff dimension is also computable (or, respectively, polynomial-time computable)? (4) If a two-dimensional curve I' is computable (or polynomial-time computable) and its Hausdorff dimension is also computable (or polynomial-time computable), does it follow that its Hausdorff measure is also computable (or, respectively, polynomial-time computable)? (5) If a function f is computable (or polynomial-time computable), does it follow that its (generalized) Julia set J ( f ) is also computable (or, respectively, polynomial-time computable)? For most of these questions, we present a negative answer. That is, through the construction of various fractal curves, we show that there exist feasibly computable curves (or functions) whose related properties are not even computable. In the following, we assume basic knowledge of computable analysis and complexity theory of real functions. For a complete treatment of these subjects, see Pour-El and Richards [13], Weihrauch 1171, KO [6] and [8]. For the basic notions of complexity classes discussed in this paper, see Du and KO [4]. 2. Computability of a Two-Dimensional Region
First, we present a brief review of the notion of computable real functions. We use Cauchy sequences of dyadic rational numbers to represent real numbers. Let D denote the set of dyadic rationals; i.e.,
Each dyadic rational d has an infinite number of binary representations. For instance, d = 318 may be represented by
0.011
,oo...o_ k
with an arbitrary number k of trailing zeroes. A binary representation s of a dyadic rational d is said to have n bits ( i n the fractional part) if s has
254
exactly n bits to the right of the binary point. We let D denote the set of binary representations of dyadic rationals, and let D, be the set of all binary representations s E D which have exactly n bits. When there is no confusion, we use s E D to denote both a binary representation in D and the dyadic rational d it represents. We say a function q5 : N + D binary converges to a real number x, or represents a real number x, if (i) for all n 2 0, q5(n) has a binary reprefor all n 2 0, )x- q5(n)I 5 2-". For sentation of at most n bits, and (ii) each function q5 that binary converges to x, there is a corresponding set representation of 2, namely, the set L6 = { s E D 1 ( I n )[s has n bits and s
< #(.)I}.
(Note that L4 is defined as a set of binary representations of dyadic rationals, instead of a set of dyadic rationals. It is possible that two different s , t E D represents the same dyadic rational d, but one is in L4 and the other is not.) We call such a set a left cut of x. For any x E R, there is a unique function q5, : N + D that binary converges to x and satisfies the condition x - 2-, 5 &(n) < x for all n 2 0. We call this function #, the standard Cauchy function for x. The left cut L#x that corresponds to the standard Cauchy function 4, for x is called the standard left cut of x, and we write L, for L # x . It is easy to see that L, = {s E D 1 s < x}. (Note that the set L, may be treated as a set of dyadic rationals. The membership of a dyadic rational d in L, is independent of its binary representations.) To compute a real-valued function f : R + R, we use oracle Turing machines as the computational model. A function f : R + R is said to be computable if there is an oracle Turing machine M , that uses a function q5 : N + D that binary converges to a real number x as an oracle, such that on input n E N, M"n) outputs a string d E D, such that Id- f(x)l 5 2-n. This notion of computable real functions can be extended to functions from R to R2 (and functions from R2 to R2) in a natural way. More precisely, a function f : R + R2 is computable if there is an oracle Turing machine M that, on oracle q5 which binary converges to z E R and an input n E N, outputs two strings d l , d2 E D n such that both [ d l-y1 I and Id2 -y21 are bounded by 2-,, where f(x) = (yl,y2). To compute a function f from R2 to R2, we modify the above oracle Turing machine M to use two oracle functions q51,q52 : N + D which binary converge to two real numbers x1 and x2. We now extend this concept of computability to two-dimensional subsets
255
S E R2. For any set S R2, let xs denote its characteristic function; i.e., x s ( ( x , y ) )= 1 if ( x , y ) E S and x s ( ( x , y ) )= 0 if ( x , y ) $ S. Intuitively, a set S R2 is computable if xs is a computable function, or, equivalently, if there is an oracle Turing machine M that uses two functions that represent a point ( x , y ) in R2 as oracles and decides whether ( x , y ) E S. However, we notice that the function xs is discontinuous at the boundary of S , and hence, from a well-known result in computable analysis, it cannot be decided completely by the Turing machine M (unless S is trivial). Instead, we must allow M to make mistakes. There are several ways to relax the requirement on machine M . Here, we present two formal definitions. The first one is recursive approximability, which is an extension of the notion of recursive approximability of sets of real numbers in KO [6]. Informally, a set S Rz is recursively approximable if there is an oracle Turing machine M that takes two functions representing a point ( x ,y ) E R2 as the oracles and an integer n as the input, and decides whether ( x ,y) is in S in R2 of Lebesgue such a way that the errors of M occur in a set E,(M) measure less than 2-n. We write p ( A ) to denote the two-dimensional Lebesgue measure of a set A 5 R2, and p * ( A ) the outer measure of set A. Definition 2.1. A set S 5 R2 is recursively approximable if there exists an oracle Turing machine M such that for any oracles (6,,$) representing a point ( z , y ) E R2 (i.e., 6, and $ binary converge to x and y, respectively), and for any input n, M outputs either 0 or 1 such that the following set E,(M) has measure p*(E,(M)) 5 2-n:
E n ( M ) = { ( x , y )E Rz I there exists ($,$) representing ( x , y ) such that M @ > @ (#n x) s ( ( x , y ) ) } . The notion of recursively approximable sets is equivalent to the notion of recursively measurable sets of Sanin [15]. The second notion of computability of a two-dimensional subset of R2 is recursive recognizability. Informally, a set S C R2 is recursively recognizable if there exists a Turing machine that recognizes whether two given oracle functions represent a point ( x , y ) in S correctly for all points that have a distance greater than 2-, from the boundary of S , where n is the input.
Definition 2.2. A set S C R2 is recursively recognizable if there exists an oracle Turing machine M that works as in Definition 2.1 such that E,(M) C_ ( ( 2 ,y) E R2 I the distance between ( x ,y) and the boundary of S is 5 2-"}.
256
Both of the above two computability notions of two-dimensional regions allow two-sided errors to occur in the computation; that is, errors may occur at some point ( X I , y1) E S and also at some point ( 2 2 , y2) # S. For certain types of regions, we might want to consider the notion of computability with only one-sided errors. For instance, if a set S has Lebesgue measure equal to zero, then it is trivially recursively approximable, since an oracle Turing machine that always outputs 0 has error measure p(E,(M)) = 0. Such an oracle Turing machine also recursively recognizes S since all the errors occur at points in S and hence at the boundary of S. To more correctly reflect the notion of computability on such sets, we would like the oracle Turing machine to have errors only when the oracles represent a point ( 2 ,y ) that is not in S . We call a recursively approximable (or recursively recognizable) set S R2 strongly recursively approximable (or, respectively, strongly recursively recognizable) if the underlying oracle Turing machine M satisfies the additional condition that E n ( M ) n S = 0. The particular type of subsets of the two-dimensional plane we are interested in here is the class of bounded simple regions, i.e., bounded, connected subsets of R2 which contain no holes. For simple regions whose boundaries are simple closed curves (i.e., Jordan curves), we have a third natural notion of computability: the computability of its boundary as a function from [0,1] to R2. That is, we may consider a region S computable if there is a computable real function f : [0,1] + R2 whose range is exactly the boundary rs of region S. The relations between these notions of computability on twodimensional regions are not simple. We first list some positive results.
Theorem 2.1. (Chou and KO [3]) (a) If the boundary of a region S is a computable Jordan curve, then S is recursively recognizable. (b) If the boundary of a region S is a computable Jordan curve of a finite length, then S is recursively approximable. (c) If S is recursively approximable, then S is recursively recognizable. The proofs of the following negative results about these computability notions are based on the fractal construction of Section 4.
Theorem 2.2. (a) There exists a computable Jordan curwe r whose interior S is not recursively approximable. (b) There exists a recursively recognizable set S that is not recursively approximable. For the more precise relation between a computable Jordan curve and
257
recursive approximability of its interior, see Theorem 4.l(b). Finally, we remark that most relations between strongly recursively approximable sets and strongly recursively recognizable sets are negative ones. We summarize these results below. Since the proofs are relatively simple and do not involve the construction of fractal curves, we omit them here.
Theorem 2.3. (Chou and KO [3]) (a) There exists a recursively approximable set that i s n o t strongly recursively approximable. (b) There exists a recursively recognizable set that is n o t strongly recursively recognizable. ( c ) T h e class of strongly recursively approximable sets and the class of strongly recursively recognizable sets are incomparable. 3. Computational Complexity of Two-Dimensional Regions
Since the notion of computable real functions is defined by oracle Turing machines, it is natural to extend this notion to polynomial-time computable real functions. Namely, we say a function f : P -+ P is polynomial-time computable i f f is computable by an oracle Turing machine M such that for all oracles 4 and all inputs n, M b ( n ) halts in time p ( n ) for some polynomial function p . Polynomial-time approximable and polynomial-time recognizable sets S P2 can also be defined based on the time complexity of oracle Turing machines.
s
Definition 3.1. (a) A set S C R2 is polynomial-time approximable if it is recursively approximable by an oracle Turing machine M that, on input n, runs in time p ( n ) for some polynomial p , regardless of what the oracles are. P2 is polynomial-time recognizable if it is recursively (b) A set S recognizable by an oracle Turing machine M that, on input n, runs in time p ( n ) for some polynomial p , regardless of what the oracles are. The notions of strongly polynomial-time approximable sets and strongly polynomial-time recognizable sets can be defined in a similar way. Strongly polynomial-time recognizable sets are interesting since they can be used to characterize the sets of zeroes of polynomial-time computable functions from [0, 112to R2. (A function f : [0, 112 -+ R2 is polynomial-time computable if it is computable by a two-oracle Turing machine M such that for all oracles 4,q!~and all inputs n, Mb!G(n) halts in time p ( n ) for some polynomial function p . ) Rettinger and Weihrauch [14] also used the notion of local t i m e complexity, which is closely related to the notion of strongly
258
polynomial-time recognizability, to study the computational complexity of Julia sets. Regarding the relations between these notions of polynomial-time computability of two-dimensional regions, we note that results in Theorem 2.1 do not extend to their polynomial-time counterpart immediately. They depend on the unknown relations between discrete complexity classes. To avoid the definitions of some nonstandard complexity classes, we only present the relations in terms of well-known complexity classes. For more details, see Chou and KO [3]. The following complexity classes of sets of strings are well known:
P : the class of sets computable in polynomial time by deterministic Turing machines; FP: the class of functions computable in polynomial time by deterministic Turing machines; BPP: the class of sets computable in polynomial time by probabilistic Turing machines, with bounded errors; NP: the class of sets computable in polynomial time by nondeterministic Turing machines; # P : the class of functions that count the number of accepting computation paths of a polynomial-time nondeterministic Turing machine; UP: the class of sets computable in polynomial time by nondeterministic Turing machines that have at most one accepting computation path.
c
c
c
It is known that P UP N P , P BPP, and F P C # P . It is also [ P = BPP]. It is known that [FP = #PI + [ P = NP], and [FP = #PI not known whether any of the above equations holds.
+
+
Theorem 3.1. (Chou and KO [3])In the following, (a) (b) + (c). (a) FP = # P . (b) All polynomial-time approximable subsets of [0,112 are polynomialt i m e recognizable. (c) BPP = P . Theorem 3.2. (Chou and KO [3])In the following, ( u ) 3 ( b ) + ( d ) and ( a ) ( 4 + (4. (a) FP = # P . (b) If r i s a polynomial-time computable Jordan curve, t h e n the interior S of I? is polynomial-time recognizable.
*
259
(c) If I? is a polynomial-time computable Jordan curve, and i f I? has a finite length, then the interior S of is polynomial-time approximable. (d) UP = P ; i.e., one-way functions do not exist. In other words, if BPP # P and UP # P , then the relations of Theorem 2.1 do not hold for their polynomial-time counterparts. To demonstrate the proof techniques for these results, we present in the following, a sketch of the proof of a weaker result of (b) (c) of Theorem 3.1.
+
Theorem 3.3. If there exists a set T C {1}*(called a tally set) in the class BPP - P , then there is a nondegenerate rectangle [a,b] x [c, d] that is polynomial-time approximable but not polynomial-time recognizable.
Sketch of Proof. The main observation is that there is a simple construction of a real number x from a given tally set T 5 {1}*such that x and T have the same time complexity, within a polynomial factor. Namely, define M
z = C(XT(10)
+ 1) 4-". *
n= 1
Then, it can be proved that x and T satisfy the following properties: (1) The standard left cut L, of z is polynomial-time computable relative to T (i.e., it can be computed in polynomial time by an oracle Turing machine that uses set T as an oracle). (2) Set T is polynomial-time computable relative to x (i.e., T can be computed in polynomial time by an oracle Turing machine that uses a function $ that binary converges to x as an oracle). Thus, for T E BPP - P , we know that x is not polynomial-time computable but its standard left cut L, is in BPP. Let A be the rectangle [0,x] x [0,1]. We claim that A satisfies the required conditions. First, we check that A is not polynomial-time recognizable. Assume, by way of contradiction, that A is polynomial-time recognizable by an oracle Turing machine M . Then, we can compute an approximation d to x within error 2-" as follows: We simulate M to binary search for a dyadic rational e E D which has a binary representation in Dn+l such that M(e*1/2)(n+l) = 1and M(e'91/z)(n+l) = 0, where e' = e+2-(n+1). Then, either M is correct on these two computations and hence e 5 x 5 el, or M makes a mistake on at least one of these two computations and hence e or el is within distance 2-("+l) of the right line segment of the rectangle. In either case, we know that ( x - el 5 2-n and ( x - e'( 5 2-". One of the numbers e or e' has
260
a binary representation in D, and is the required approximation. This contradicts the fact that x is not polynomial-time computable. Conversely, we check that A is polynomial-time approximable. Let M I be a probabilistic Turing machine that accepts the standard left cut L, of x in polynomial time. It means that there is a two-input deterministic Turing machine and a polynomial function p such that, for all dyadic rationals d c Dn7 (a) if d < x then Prlwl=p(n)[al(d, w) = 11 > 1 - 2-,, and (b) if d > x then Prlwl=p(n)[ul(d,w)= 11 < 2-". (In the above, we have used a standard method of amplification of accepting probability on probabilistic Turing machine M I ;see Theorem 8.19 of Du and [41.) Now,we design an oracle Turing machine M as follows: For any oracle functions 47II, that represent a point ( y ,z ) in [0,112, and for any input n, it first gets d, = d ( n + 3 + p ( n + 3 ) ) and d, = II,(n+3). Then, it decodes dy to d and w where d consists of the first n 3 bits of d, and w is the string of the last p ( n 3 ) bits of d, (i.e., in the binary expansion form, d, = 0 . d ~ ) . M accepts if and only if (d, w) = 1 and 0 5 d, 5 1. It is not hard to verify that M only makes mistake at a point ( y ,z ) if (1) lz - 11 I 2-("+3) or Iz - 01 < - 2-("+3), (2) ly - X I 5 2-("+3) or 1y - 01 5 2-(n+3), ( 3 ) d, = 0.dw a n d d < x b u t ~ 1 ( d , w ) = O o r ( 4 ) d Y = O . d w a n d d > x b uut l ( d , w ) = l . The outer measure of these error areas is, for each of (1) and ( 2 ) , at most 2-(n+2) and, for ( 3 ) and (4) together, at most 2-("+2). Altogether, we have p * ( & ( M ) ) I 2-,. 0 KO
+
+
The negative results of Theorem 2.2 can be extended to the polynomialtime setting.
Theorem 3.4. (a) There exists a polynomial-time computable Jordan curve I? whose interior S is not recursively approximable. (b) There exists a polynomial-time recognizable set S 5 R2 that is not recursively approximable. For strongly polynomial-time computability notions of two-dimensional regions, all negative results of Theorem 2.3 still hold.
Theorem 3.5. [Chou and KO, 31 (a) There exists a polynomial-time approximable set that is not strongly polynomial-time approximable. (b) There exists a polynomial-time recognizable set that is not strongly polynomial-time recognizable.
26 1
(c) The class of strongly polynomial-time approximabke sets and the class of strongly polynomial-time recognizable sets are incomparable.
4. Fractal Curves and Their Interiors
In this section, we consider Questions (1) and (2) listed in Section 1. We first give a complete characterization of the computability of the interior of a Jordan curve in terms of the Lebesgue measure of the curve itself. Recall that a set S E R2 is r.e. open if there exist recursive functions 4, $, 0 : N + D such that S = Ur='=, R,, where R, is the 0(n) x O(n) open square centered at ( $ ( n ) , $ ( n ) )We . say T R2 is r.e. closed if R2 - T is r.e. open. A real number x is called left r.e. if its standard left cut L, is r.e.; it is called right r.e. if -x is r.e. The relations between the computability of a Jordan curve r and the computability of its interior S and its measure p ( S ) can be summarized as follows: Theorem 4.1. (KO and Weihrauch [lo]) (a) If a Jordan curve r is computable, then its interior S and its exterior T are r.e. open sets. Furthermore, the measure p ( S ) must be a left r.e. real number, and the twodimensional Lebesgue measure p ( r ) must be a right r.e. real number. (b) If a Jordan curve is computable and the two-dimensional Lebesgue measure p ( r ) of the curve itself is zero, then its interior S is recursively approximable. (c) If a Jordan curve I? is computable and the two-dimensional Lebesgue measure p ( F ) of the curve itself is a recursive real number, then the measure p ( S ) of its interior S must be a recursive real.
It is interesting to note that, in Theorem 4.l(b) and 4.l(c), the curve may be a fractal (a fractal curve may have measure zero in the twodimensional plane). That is, even if a curve is a fractal, its interior is still recursively approximable-as long as the two-dimensional measure of I' is zero. On the other hand, if the curve r has a nonrecursive two-dimensional measure (and hence must be a fractal), then it may have a noncomputable interior.
r
Theorem 4.2. (KO and Weihrauch [lo]) For any left r.e. real number x > 0 , there is a polynomial-time computable Jordan curve r whose interior has Lebesgue measure equal to x.
262
Figure A basic curve of rn. The dash line is a basic line segment c r,-1, ant the dot lines are the basic curves of rn+l that replaces the basic line segments of r,.
Proof Idea. The proof of this result is based on the construction of a monster curve. (A curve is called a monster curve if it is a simple curve and yet its two-dimensional measure is greater than zero [ll].)It is a standard iterative construction that is used to define, for instance, the Koch curve. Namely, we start with a line segment ro. Then we construct at each stage n > 0, a curve ,?I and let r be the limit of I?,. Each curve I?, has 4n basic line segments, and I'n is constructed from rn-l by replacing each basic line segment of r,-I by a basic curve that contains four shorter basic line segments (see Figure 1). By choosing a suitable initial basic line segment I'o, we can make the limit curve I? of I?, to have a positive measure q for any given dyadic rational q > 0. Correspondingly, the interior of r has a measure r E ED. Intuitively, if x is left r.e., then there is a Turing machine M that outputs an increasing sequence dl ,d z , . . . of dyadic rationals that converges to x. If we simulate this machine for i moves, and output the largest dj that M generates in i moves, this sequence can be generated by a polynomial-time Turing machine. Formally, if x is left r.e., then there exists a polynomialt i m e computable function e : N + N such that e ( i ) 2i and
<
i=O
where sgn(m) = 1 if m 2 1 and sgn(m) = 0 otherwise.
263
To prove the theorem, we modify the above construction of as follows: At stage n, we select 22"-e(") basic line segments of r,. On each basic line segment we selected, we replace the line segment with a suitable curve which does not include any basic line segment. On each remaining basic line segment, we replace the line segment with a basic curve of rn+l. In the original construction of the monster curve r, the limit of the basic curves of I?,, rn+l,.. . has a positive measure q/22n around each basic line segment of I?,. Thus, the interior of r loses measure q/22n to its boundary around this basic line segment. Since rn has 22n basic line segments, the interior of totally loses measure q. Now, our modification reduces the number of basic line segments in I?, from 22n to 22n-e(") so this modification gains measure q . 2-e(n) for the interior. In the limit, we gain measure q x for the interior, where x = s g n ( e ( n ) ).2-e(n). Thus, the measure of the interior of the limit curve becomes r + qx. 0 1
Since there exist left r.e. real numbers that are not recursive, and since the Lebesgue measure of a recursively approximable region in R2 must be a recursive real [3], we have the following negative result regarding the first half of Question ( 2 ) of Section 1. This also proves Theorems 2.2 and 3.4.
Corollary 4.1. There exists a polynomial-time computable Jordan curve whose interior has a nonrecursive two-dimensional Lebesgue measure. For the second half of Question ( 2 ) , we can show a stronger result: For a polynomial-time computable Jordan curve, its interior does not have to be polynomial-time approximable, even if the Lebesgue measure of the curve itself is zero. The proof is based on a similar iterative construction of a curve F, with the modification that the Hausdorff dimension of the curve is actually equal t o one (and hence the curve is not a fractal) but the length of the curve is still infinite. (We delay the formal definition of Hausdorff dimensions until Section 5.)
Theorem 4.3. (KO and Weihrauch [lo]) For any recursive real number x > 0 , there is a polynomial-time computable Jordan curve r whose twodimensional Lebesgue measure is zero and whose Hausdorff dimension is equal to one, such that its interior has Lebesgue measure equal t o x . Corollary 4.2. (a) There exists a polynomial-time computable Jordan curue r whose two-dimensional Lebesgue measure is zero and whose Hausdorff dimension is equal to one, such that its interior has a measure that is not polynomial-time computable.
264
(b) There exists a polynomial-time computable Jordan curve I? whose two-dimensional Lebesgue measure is zero and whose Hausdorff dimension is equal to one, such that its interior S is not polynomial-time approximable. Note that the length of the curve I? in Theorem 4.3 and Corollary 4.2 is infinite. This property of curve I? is critical. In fact, if a Jordan curve I? is of a finite length, then the Lebesgue measure of its interior must be polynomial-time computable relative to a function in # P .
Theorem 4.4. The following are equivalent: (a) Each function f : {1}*-+ N that belongs to # P is actually in FP. (b) For each polynomial-time approximable set S [0,112, the measure p(S) is polynomial-time computable. (c) If a Jordan curve I? is polynomial-time computable and has a finite length, then the measure p(S) of its interior S is also polynomial-time computable. Finally, we note that a simple modification of the construction of Theorem 4.2 yields a negative result for Question (1) of Section 1.
Theorem 4.5. (KO [7]) There exists a polynomial-time computable Jordan curve I? that has finite length, but its length is a left r.e., nonrecursive real number. 5. Hausdorff Dimensions and Hausdorff Measure
A fractal is usually defined as a set whose Hausdorff dimension is strictly higher than its topological dimension (see Mandelbrot [ll]).First let us review the notion of Hausdorff dimensions. Let A be a subset of R". A S-covering of A is a countable collection {Ui}g1of sets in R" such that A Ui and IUi( 5 6, for i 2 1, where lUil = sup{lx - yl I x , y E Ui}. For any real number s > 0, the s-dimensional Hausdorff measure of A is defined as
Ugl
xtl"(A)= liminf 6-0
{ 2 IVils I {Ui}zl
is a &-coveringof A
i=l
1.
It is not hard to verify that Ns(A) is nonincreasing with respect to s. The Hausdorff dimension of set A is formally defined to be dimH(A) = inf{t E R I Nt(A) = 0). We say a subset A R is a fractal if 0 < dimH(A) < 1; and a curve the two-dimensional plane is a fractal if dirnH(I7) > 1.
r on
265
The above definitions of Hausdorff measure and Hausdorff dimension are quite abstract and do not seem to provide a natural method for calculation. As a consequence, the computation of the Hausdorff dimension of a set is considered very difficult. In the following, we justify this intuition on fractal sets in the one-dimensional line R and fractal curves on the two-dimensional plane Rz. First recall that a set S R is recursively open if there exist recursive functions 4, $ : N + D and 0 : N + N such that (i) S = u,"==,($(n), $(n)), and (ii) p ( U ~ B ( n ) ( ~ ( k ) , $ ( k<) 2-,; ) ) that is, it is r.e. open and we can approximate S recursively. A set T 5 R is recursively closed if R - T is recursively open. For a subset of R, we note that if its Hausdorff dimension is less than 1, then its Lebesgue measure is zero and hence recursively approximable. So, it is natural t o consider strongly polynomial-time approximable sets.
c
T h e o r e m 5.1. (KO [9]) There exists a recursively closed, strongly polynomial-time approximable set S R such that dimH S is a nonrecursive real number between 0 and 1.
c
This result can be extended to two-dimensional curves.
Theorem 5.2. (KO [9]) There exists a polynomial-time computable function f : [0,1] + Rz that defines a simple curve r such that dirnHr is a nonrecursive real number between 1 and 2. Proof Idea. We start with a left r.e., nonrecursive real number a= X4-k, kEK
where K is an r.e., nonrecursive subset of N. Similar to the proof of Theorem 4.2, we can find a polynomial-time computable sequence{a,} of dyadic rationals that converges to a, with the property of an 5 an+I < a, for all n 3 0. Next, we construct a simple fractal curve I? on R2 whose Hausdorff dimension is equal to 3/2. This curve r is the limit of curves I',. I'o is a horizontal line segment of length 1, which is a basic line segment. The curve rn consists of 8, basic line segments, each of length 4-,. To get curve rn+l,we replace each basic line segment of I?, by a basic curve that consists of 8 shorter basic line segments (see Figure 2). To prove the theorem, we modify the curves ,?I to A, as follows: For each n, define a number p , such that 2, < p , 5 2n+1. Let 111 = rl.
266
Figure 2.
A basic curve of
rn changing to 8 basic curves of rn+l.
Assume that the curve A2n has been defined, with m2n basic line segments. For all k such that 2, < k 5 2, + p , , we replace each basic line segment of A k - l by a basic curve as shown in Figure 2. For all k such that 2, + p , < k 5 2,+l, we let A k = (i.e., each basic line segment of A k - 1 becomes 4 basic line segments of A k ) . By choosing suitable values for p , (depending on a , and a,+l), we can make m2n to be exactly 82n(1-an). Now, we claim that the limit curve A of A, has dimH(A) = (3/2)(1-a), which is a nonrecursive real number. Since the curve A is the limit of simple curves A,, it can be seen that the Hausdorff dimension of A is equal to its box-counting dimension (see Peitgen et al. [12] and Falconer [ 5 ] ) . To cover A with 4Kk x 4-k boxes, we only need m k boxes (each of the diamond shape 0 with a basic line segment of h k as one of its diagonals). Thus, the box-counting dimension of A is equal to lim
k+cc
log m
-
k
= lim
log 82"( l - a n - 10g4-~"
= lim
3.27yl - a n ) 3 = - . (1 - a ) . 2.2, 2
n+cc
n+cc
0
Intuitively, the Hausdorff dimension provides a measure of the relative density of a fractal set. Among all sets of the same Hausdorff dimension, Hausdorff measure provides a finer comparison of their size. For a set A C B with the Hausdorff dimension equal to one, its Hausdorff measure (with respect t o dimension one) is equal to a computable constant times its Lebesgue measure. Therefore, if the Hausdorff dimension of a recursively approximable set A is one, then its one-dimensional Hausdorff measure is either infinite or is a finite recursive real number. However, when the Hausdorff dimension s of A is a fraction, then the s-dimensional Hausdorff measure of A may not be computable.
Theorem 5.3. (KO [9]) There exists a recursively closed, strongly polynomial-time approximable set T 5 [0,1] such that dimH T = 1/2 and
267
its (1/2)-dimensionalHausdorff measure 7-11/2 ( T ) is a finite, nonrecursive real number. 6. Julia Sets
An important group of fractals is generated by dynamical systems. In this section, we consider Julia sets of computable real functions f defined on the one-dimensional real line. Let f : R + IW be a real-valued function. We let f denote the iteration of f for n times; i.e., f O(x) = x and Jn+’ (x)= f(f”(x)). We define the (generalized) Julia set o f f : B + B to be J ( f ) = {x E
I (3B E B) (Vn) Ifn(x)l 5 B ) .
Cenzer [2] gave a characterization of Julia sets of computable onedimensional functions.
Theorem 6.1. (Cenzer [2]) If A C R is a bounded, r.e. closed set with either a computable maximum point or a computable minimum point, then there is a polynomial-time computable function g such that J ( g ) = A . This characterization shows that most r.e. closed sets can be Julia sets and, hence, from the results of Section 4,that there exist easily computable functions f whose Julia sets are hard to compute.
Corollary 6.1. (a) There exists a polynomial-time computable function + B such that J (f) is not recursively approximable. (b) There exists a polynomial-time computable function f : E% + R such that the Hausdorff dimension of J ( f ) is not a recursive real number.
f : E%
It is well known that, for any bounded r.e. closed set A , there exists a computable function h such that h ( z ) 2 0 for all x E Iw and A = {x E B I h ( z )= 0 ) . The main idea of the proof of Cenzer’s theorem is to modify this function into a new function g so that if h(x) > 0 then limn-,m gn (x)= ca and hence 2 !$ J ( g ) and if h(x) = 0 then g k ( z )= 0 and x E J ( g ) . Thus, deciding whether a point x is in J ( g ) or not is equivalent to deciding whether h(x) > 0, and the theorem follows from the fact that deciding whether two given Cauchy functions represent two distinct real numbers or not is undecidable (see KO [6]). Intuitively, we suspect that a Julia set of a computable function f is difficult to compute because of the unpredictability of the behavior of fk(x) after a large number of iterations, rather than the inability to determine whether f(x) > 0. To understand more clearly why Julia sets are hard to
268
compute, we show in the following that a dynamical system can actually simulate the computation of a universal Turing machine M such that its Julia set corresponds to the complement of the halting set of M .
Theorem 6.2. (KO [9]) Let M be a Turing machine over the alphabet C. There exist polynomial-time computable functions f : [0,1] -+ R and h : C* + R such that, for any w E C* of length n, (i) i f M accepts w , then [ h ( w ) ,h ( w ) cVn]n J ( f ) = 8, and (ii) i f M does not accept w , then [ h ( w ) , h ( w ) c - ~ C ] J(f), where c is a positive constant.
+
+
Sketch of Proof. Basically, we treat the computation of M as a discrete dynamical system. That is, the mapping from a machine configuration to its successor configuration is a polynomial-time computable function g , and the iteration of this function is a dynamical system whose behavior is equivalent to the computation of the Turing machine M . The function f in the theorem is, thus, just a continuous function that encodes the discrete function g , and h a function that on input w encodes the halting configuration of M on w . 0 Corollary 6.2. (KO [9]) There exists a polynomial-time comuptable function f : R -+ R such that J ( f ) is not recursively approximable. I n addition, f satisfies the following property: f o r any two dyadic rationals d , e , the question of whether f ( d ) > e is decidable in polynomial time.
Finally, we remark that there are studies of Julia sets of some specific dynamical systems. For instance, Rettinger and Weihrauch [14] have studied the Julia sets defined by the one-complex-variable functions of the form f(z) = z2 c. Other related results can be found in Saupe [16] and Zhong P81.
+
References 1. Barnsley, M. F., Fractals Everywhere, Academic Press, 2nd Edition, Boston, 1993. 2. Cenzer, D., Effective real dynamics, in Logical Methods, J. N . Crossley, J. B. Remmel, R.A. Shore, M. E. Sweedler, eds., Birkhauser, Boston, 162177, 1993. 3. Chou, A. W. and KO, K., Computational complexity of two-dimensional regions, SIAM J. Comput. 24, 923-947 (1995). 4. Du, D.-Z. and KO,K., Theory of Computational Complezity, Wiley, New York, 2000.
269
5. Falconer, K., Fractal Geometry, Mathematical Foundations and Applications, Wiley, New York, 1991. 6. KO, K., Complexity Theory of Real Functions, Birkhauser, Boston, 1991. 7. KO,K., A polynomial-time computable curve whose interior has a nonrecursive measure, Theoret. Comput. Sci. 145,241-270 (1995). 8. KO, K., Polynomial-time computability in analysis, in Handbook of Recursive Mathematics, Vol. 2: Recursive Algebra, Analysis and Combinatorics, Yu. L. Ershov, S.S. Goncharov, A. Nerode and J. Remmel, eds., Studies in Logic and the Foundations of Mathematics, Vol. 139, Elsevier, Amsterdam, pp. 1271-1317, 1998. 9. KO, K., On the computability of fractal dimensions and Hausdorff measure, Annals of Pure and Appl. Logic 93,195-216 (1998). 10. KO, K. and Weihrauch, K., On the measure of two-dimensional regions with polynomial-time computable boundaries, in Proceedings of I 1 th IEEE Conference on Computational Complexity, IEEE Computer Society Press, 150-159, 1996. 11. Mandelbrot, B., The Fractal Geometry of Nature, W. H. Freeman, New York, 1983. 12. Peitgen, H.-O., J ~ r g e n sH. , and Saupe, D., Chaos and Fractals, New Frontiers of Science, Springer-Verlag, New York, 1992. 13. Pour-El, M. and Richards, I., Computability in Analysis and Physics, Springer-Verlag, Berlin, 1989. 14. Rettinger, R. and Weihrauch, K., The computational complexity of some Julia sets, in Proceedings of the Fifth Workshop on Computability and Complexity in Analysis, V. Brattka, M. Schroder, and K. Weihrauch, eds., FernUniversitat, Hagen, 159-169, 2002. 15. Sanin, N., Constructive Real Numbers and Function Spaces, Translations of Mathematical Monographs, 21,translated by E. Mendelson, American Mathematical Society, Providence, RI, 1968. 16. Saupe, D., Efficient computation of Julia sets and their fractal dimension, Physica 28D, 358-370 (1987). 17. Weihrauch, K., Computable Analysis, A n Introduction, Springer-Verlag, Berlin, 2000. 18. Zhong, N., Recursively enumerable subsets of Rq in two computing models, Blum-Shub-Smale machine and Turing machine, Theoret. Comput. Sci. 197, 79-94 (1998).
270
DEFINABILITY IN LOCAL DEGREE STRUCTURES - A SURVEY OF RECENT RESULTS RELATED TO JUMP CLASSES
ANGSHENG LI* Institute of Software, Chinese Academy of Sciences, P.O. Box 8718, Beajing, 100080, P. R. of China. E-mail: angshengaios. ac. cn
YUE YANG+ Department of Mathematics, Faculty of Science, National University of Singapore, Lower Kent Ridge Road, Singapore 119260. E-mail: mat yangy aleonis. nus. edu. sg
Keywords: Computably enumerable degrees, jump classes, definability, ideals, n-c.e. degrees We review recent developments in the study of local degree structures, with emphasis on relations which are definable by local properties.
1. Introduction The study of degree structures is one of the main areas of computability theory. In this survey paper, we shall focus on Turing degrees below O f , *Partially supported by NSF Grant No. 69973048 NSF Major Grant No. 19931020 (P. R. CHINA) and by NUS Grant No. R-146-000-028-112 (Singapore). $Partially supported by NUS Grant No. R-146-000-028-112 (Singapore). Both authors would like to thank Chong Chi-tat and Wu Guohua for suggestions.
27 1
more precisely, we shall discuss properties related to jump classes in both c.e. degrees and in the difference hierarchy, i.e., in n-c.e. degrees for n > 1. In the past decade, many breakthroughs were made and some longstanding open problems were solved, notably Shore and Slaman’s result of definability of jump operators [93], and Nies, Shore and Slaman’s definability results concerning “most of’’ the jump hierarchies [79]. The methods involved, such as forcing and coding standard models into c.e. degrees, have strong “global” flavor. These methods are general and yield quite powerful results, but the definability results obtained do not have the flavour of “natural” definability. The word “natural” is meant for logically simple and coming from normal properties such as join or meet. Finding a natural definition of any of the jump classes would illuminate the role of the Turing jump in R,and a positive solution of the problem of a natural definable degree would challenge existing priority methods. The elusiveness of natural definitions suggests the lack of understanding of various aspects of the structure. However, it should be pointed out that it is not entirely clear if such a natural definition exists at all.
In any case, the significance of a natural definition is much greater than merely providing another definition of the relevant jump class. It is our hope that knowing the global definability will stimulate more efforts on the study of local properties. We also feel that the study of local properties would offer some technical breakthroughs to tackle more difficult problems. Historical examples such as Lachlan’s proof of the Nonsplitting Theorem which introduced the 0”‘-priority argument, would seem to support this hypothesis. We only focus on very selective topics in the area. There are a number of other recent important developments, and we refer to [15] and [42]. Also there can be found comprehensive lists of open problems. We especially recommend the articles by Arslanov [8], Lerman [61], Nies [78], and Shore [go] in [15] and Cooper [21], Slaman [94] and Shore [89] in [42]. For those who are interested in local Turing degree theory, we also recommend a recent survey paper by Li [65], which had reviewed many results and discussed many foundamental problems. This paper is divided into two natural parts: one on computably enumerable degrees and one on the difference hierarchy.
272
2. Cornputably Enumerable Degrees
Let R denote the structure (R,ST),where R is the set of all computably enumerable (c.e.) degrees and
+
Although the study of individual jump classes, such as high c.e. degrees, began in the 1960's, it is not until early 1970's that the notion of the general jump classes was formulated. Cooper, in the original version of [16], and Soare [95] defined the important hierarchy of the (c.e.) degrees.
Definition 2.1. The jump classes (or the high/Zow hierarchy) are defined by H, = {a E R I a(,) = O("+l)},
L, = {a E R 1 a(,) = o(,)) where n 2 0, x(,+') is the Turing jump of x("),and x(O)= x for each x. An element of H, (L,) is called high, (low,), and for n = 1, we also call an element of HI (LI) high (low). We (very) loosely group the material into the following categories: 0 0
0
Splitting and nonsplitting. Cappable and noncappable degrees. Cuppable and noncuppable degrees. Definable ideals.
273
The theme of jump classes will appear throughout. 2.1. Splitting and Nonsplitting
We say that a degree a splits if there are degrees ao,al < a such that a = a0 V al. This seemingly insignificant notion turned out to be one of the driving forces in the development of degree theory. As we shall see in the next section, we are still far from fully understanding the phenomena of splitting and nonsplitting in n-c.e. degrees.
2.1.1. Basic Results We begin with two fundamental theorems proved by Sacks in the 1960’s, namely Sacks Splitting Theorem [84] and Sacks Density Theorem [85].
Theorem 2.1. (Sacks Splitting Theorem) For any nonzero c.e. degree a, there are nonzero c.e. degrees ao,a1 < a such that a0 V a1 = a. Theorem 2.2. (Sacks Density Theorem) Given any c.e. degrees b < a, there is a c.e. degree c such that b < c < a. It is natural to ask if the two theorems above can be combined: Given any c.e. degrees b < a, are there c.e. degrees b < ao,al < a such that a0 V a1 = a? Robinson [82] showed that it is the case when b is low:
Theorem 2.3. (Robinson Low Splitting Theorem) For any c.e. degrees 1 < a, i f 1 is low, then there are c.e. degrees ao, al such that l
Theorem 2.4. (Lachlan Nonsplitting Theorem) There exist c.e. degrees b < a such that a is n o t splittable over b, namely, f o r any c.e. degrees x, y, if b < x, y < a, then x V y # a. Lachlan’s result had a much greater impact than merely settling a foundamental question. In the proof of Theorem 2.4, Lachlan introduced an extremely powerful method, now categorized as O”’-priority method. We shall refer t o the c.e. degree b in Theorem 2.4 as a Lachlan nonsplitting base and the pair b < a a Lachlan nonsplitting pair.
274
Harrington [45]extended Theorem 2.4:
Theorem 2.5. (Harrington) There exists an incomplete c.e. degree b such that 0’ is not splittable over b. Such a b is called a Harrington nonsplitting base.
It was further generalized by Leonhardi [60]: Theorem 2.6. (Leonhardi) Given n 2 1, there exists a c.e. degree d such that the interval [d,O’]c R admits an embedding of the n-atom Boolean algebra preserving the (least and) greatest element, but also such that there is no (n 1)-tuple of pairwise incomparable c.e. degrees above d which pairwise join to 0’.
+
2.1.2. Distribution of Nonsplitting Pairs and Bases
The relationship between nonsplitting bases and jump classes was first studied by Shore and Slaman (unpublished):
Theorem 2.7. (Shore and Slaman) Everg high c.e. degree bounds a Lachlan nonsplitting base. Corollary 2.1. The class of Harrington nonsplitting bases is properly contained in the class of Lachlan nonsplitting bases. Proof. Let h be a high degree over which 0’ splits. Such a degree exists, for instance, take a high diamond base (see Theorem 2.28). Then h does not bound any Harrington splitting bases. By Theorem 2.7, the corollary follows. On the other hand, Cooper, Li and Yi [30] have shown that it is not the case for other jump classes.
Theorem 2.8. (Cooper, Li and Yi) There exists a nonlow2 c.e. degree a which bounds no Lachlan (hence no Harrington) nonsplitting base. Moreover, the proof can be modified to make such an a highs If we go further down to low2 degrees, then the Lachlan nonsplitting pair disappears. This was established by Harrington (unpublished) and independently Bickford and Mills [12] (see Shore and Slaman [91]):
275
Theorem 2.9. Every low, c.e. degree bounds n o Lachlan nonsplitting pair.
Shore and Slaman (see Shore and Slaman [92]) have observed that the existing technical resources could show that low2 can not be replaced by low3: Theorem 2.10. There exists a Lachlan nonsplitting pair a is low3.
< 1 such that 1
Combining Theorems 2.9 and 2.10, we have the interesting corollary: Corollary 2.2. Th(L2) # Th(L,) ( n > 2), i.e., 10% are not elementarily equivalent.
and low, ( n > 2)
We now return to the distributions of nonsplitting bases. By the Robinson Low Splitting Theorem (Theorem 2.3), there is no low Lachlan nonsplitting base. In fact, Arslanov, Cooper and Li [9] have obtained a stronger version of low splitting theorem: Theorem 2.11. (Generalized Low Splitting Theorem) For any c.e. set A, any A: set L , if L
(1) X o u X 1 = A , X o n x , = 0 , (2) for each i E (0, l}, A $T Xi @ L . On the other hand, Cooper and Li [25] showed that the usual nonsplitting construction can be adapted to provide a low3 Harrington nonsplitting base. Finally, Cooper and Li [26] obtained the following result, which pushes the nonsplitting base to its limit. Theorem 2.12. (Cooper and Li) There exists a 10% Harrington nonsplitting base.
The proof technique of Theorem 2.12 is surprising. Roughly speaking, building a low3 Harrington nonsplitting base involves a 0"'-priority argument in a 0"'-priority environment, whereas the proof of Theorem 2.12 requires performing a 0"'-priority argument in a 0"-priority environment. We would like to raise the following problem: Question 2.1. Characterize Harrington nonsplitting bases.
276
2.2. Cappable and Noncappable Degrees
We say a degree a caps if there is a degree b such that a A b = 0, where a A b is the greatest lower bound of a and b. Equivalently, we can use more algebraic terms to say that a caps if a has a complement with respect to A. In this subsection, we survey the results related to capping. However, the topic on lattice embeddings is omitted. The interested reader is referedr to Lerman [61] and Shore [89]. 2.2.1. Basic Results After the density of R is known, Shoenfield [88] suggested an algebraic approach to understand the structure R. Conjecture 2.1. (Shoenfield) Given finite partial orderings P E Q , then any embedding of P into R, in which degree theoretic joins and 0 , 0’ are consistent with Q , can be extended to a n embedding of Q into R. In particular, Shoenfield listed two consequences of the conjecture: C1. For any c.e. degrees a,b, if a,b are incomparable, then the greatest lower bound of a, b does not exist. C2. For any c.e. degrees 0 < b < c, there is a c.e. degree a < c such that a V b equals c . In fact, Shoenfield’s Conjecture would imply that the structure R is homogeneous, hence, there are no nontrivial definable relations. Conjecture 2.1 was refuted by the minimal pair theorem, which offered a direct counterexample of statement C1, shown by Lachlan [53] and independently Yates [99]. Although refuted, Shoenfield’s Conjecture (or the refutation of it) has has significant influence in the study of almost all degree structures in computability theory. Theorem 2.13. There exist nonzero c.e. degrees a, b such that a A b = 0 . We shall refer the degrees a and b in Theorem 2.13 as a minimal pair. Theorem 2.13 introduced the meet operator A and initiated the investigation of lattice embeddings. However, Lachlan [53] and Yates [99] also showed that the meet operator A is not defined everywhere. Theorem 2.14. (Lachlan and Yates) R is not a lattice.
277
2.2.2. Cappable Degrees and J u m p Classes The existence of minimal pairs provides more instances of close relationships between structural properties and jump classes. The first instance is offered by Cooper [17]. Theorem 2.15. (Cooper) E v e r y high c.e. degree h bounds a m i n i m a l pair a, b. In fact the minimal pair a,b in Theorem 2.15 can be chosen high. Lachlan [56] proved that the highness condition in Theorem 2.15 is necessary. Theorem 2.16. (Lachlan Nonbounding Theorem) There exists a c.e. degree a # 0 which bounds n o m i n i m a l pairs.
Degree a in Theorem 2.16 is called a nonbounding degree. Theorem 2.16 is improved further by Downey, Lempp and Shore [36]: Theorem 2.17. (Downey, L e m p p and Shore) T h e r e i s a high2 n o n bounding degree. In contrast to the minimal pair theorem, Yates [loo] showed: Theorem 2.18. (Yates [loo]) There exists a c.e. degree a # 0 s u c h t h a t for a n y c.e. degree x, a A x = 0 if and only i f x = 0.
By Theorems 2.13 and 2.18, the following two important notions are introduced naturally: Say a c.e. degree a is cappable, if it is half of a minimal pair, i.e., there is a c.e. degree b # 0 such that a A b = 0, and noncappable, otherwise. We use M and NC to denote the set of all cappable and noncappable c.e. degrees, respectively. The study of cappable and noncappable degrees leads to some elegant characterizations and descriptions of subclasses of the c.e. degrees.
Before we discuss noncappable degrees, we mention the following result by Shore. We shall consider its dual problem later.
Theorem 2.19. (Shore) If a is cappable, t h e n there is a high b s w h t h a t aAb=0.
278
2.2.3. Noncappable Degrees Theorem 2.18 was extended later by Ambos-Spies [2]:
Theorem 2.20. (Ambos-Spies) There exists a nonzero c.e. degree a such that f o r any c.e. degree x, a A x exists i f and only i f a and x are comparable. The degree a in Theorem 2.20 is called a strongly noncappable (s.n.c.) degree. In fact, Ambos-Spies showed that 0' can be split into two low s.n.c. degrees. Lempp [58] extended the result further:
Theorem 2.21. (Lempp) There is a high strongly noncappable degree. Lempp [58] asked whether for each degree c, c.e.a. in 0', there is a strongly noncappable degree a such that a' = c. Lempp's question is still open. The noncappable degrees were very well understood through a dynamic notion of the promptly simple sets which was introduced by Maass [74].
Definition 2.2.
(i) A coinfinite c.e. set A is promptly simple if there is a computable function p and a computable enumeration {A,}sEwof A such that for every e , if We is infinite, then there are s and x such that x E We,at s
nA p ( a ) .
(ii) A c.e. degree a is called promptly simple, if it contains a promptly simple set. The notion of prompt simplicity leads to a beautiful algebraic decomposition of c.e. degrees:
Theorem 2.22. (Ambos-Spies, Jockusch, Shore and Soare [4]) Let PS be the set of all promptly sample degrees and LC be the set of all lowcuppable degrees, namely, the degrees which join to 0' with some low c.e. degrees. Then: (i) NC = PS = LC. (ii) M is an ideal of R. (Recall that I is a n ideal of R if I is closed under join and I is closed downwards.) (iii) NC is a strong filter in R, i.e., it is closed upward in R, and if a, b E NC, then there is a c E NC with c 5 a, b.
279
Clearly, the ideal M is defined naturally, which gives us one of the only two known examples of ideals of R which are naturally defined (the other one consisting of all noncuppable degrees which we discuss later). In addition, it is easy to see that every noncappable degree a bounds a minimal pair: Simply take a minimal pair x , y and take x1 (and y1) to be any nonzero degree which is below x and a (and y and a, respectively). Such x1 and y1 exist since a is noncappable. Then x1 and y1 is a minimal pair below a. Consequently, every nonbounding degree b is cappable, in fact, for any minimal pair x, y , either x A b = 0 or y A b = 0; and every nonzero c.e. degree bounds a nonzero cappable degree. However Li [64] found an interesting fact about cappable and noncappable degrees. Theorem 2.23. (Li) There exist c.e. degrees b bound the same class of cappable degrees.
< a such that a and b
2.2.4. Continuity of Capping Historically, the continuity problem originated from Lachlan’s major subdegree problem, which we shall discuss later. The following result by Harrington and Soare 1461 deals with the dual problem of major subdegree. Theorem 2.24. (Continuity of Capping Theorem) For any c.e. degrees a , b # 0 , if aA b = 0, then there exists a c.e. degree c > a such that
cAb=O. Theorem 2.24 was then extended by Seetapun [86], giving a full solution to the dual of the major subdegree problem (see Giorgi [41]). Theorem 2.25. (Seetapun Continuity Theorem) For any c.e. degree b # O,O’, there exists a c.e. degree a > b such that for any c.e. degree x, b A x = 0 if and only if a A x = 0 .
A consequence of Theorem 2.25 is the continuity of Lachlan nonbounding degrees: Corollary 2.3. There is n o maximal Lachlan nonbounding degree. Proof. Let b be any nonbounding degree and a be as in Theorem 2.25. We show that a is also nonbounding by contradiction. Let x and y be a minimal pair below a. By the discussion before Theorem 2.23 either x A b = 0 or
280
y A b = 0. By choice of a we have either x A a = 0 or y A a = 0, we have the desired contradiction. 0
2.2.5. Diamond Bases Next we turn to the topic of diamond bases, which is related to both joining to 0‘ and meeting to 0.
Definition 2.3. We say that a c.e. degree a is a diamond base, if a is the bottom of a diamond of c.e. degrees with top 0‘, i.e., there is a degree b such that aV b = 0‘ and aA b = 0. Let DB be the set of all diamond bases. Lachlan’s Nondiamond Theorem [53]not only discovered some connections between the Turing jump and the operators V and A in R,but also had a big impact on lattice embedding problems, which is beyond this survey paper.
Theorem 2.26. (Lachlan Nondiamond Theorem) There are n o c.e. degrees a,b such that 0 < a < 0’, aV b = 0’ and a Ab = 0. Theorem 2.26 showed that 0 # DB. More connections between diamond bases and jump classes are known. The first one is also observed by Lachlan (see Ambos-Spies [l]).
Theorem 2.27. DBnL, = diamond base.
0, that is, every low c.e. degree bounds n o
Theorem 2.27 cannot be improved, as one can construct a low2 diamond base, by combining Fejer’s branching degree construction ([38]) with splitting . On the high degrees, Li and D. Yang [70] have the following result.
Theorem 2.28. DB n HI
# 0, that is,
there is a high diamond base.
Li, Slaman and Y . Yang [67] studied further on the distribution of diamond bases:
Theorem 2.29. (Li, Slaman and Y . Yang) There exists a n o n l o w c.e. degree which bounds n o diamond base.
28 1
We propose the following questions on diamond bases: Question 2.2.
(i) Does every high degree bound a diamond base? (ii) Does DB have minimal elements? Settling this problem will either give us more continuity of the c.e. degrees, or more definable antichains of R. (iii) Characterize the c.e. degrees which bound elements of DB. Characterize the atomic jump classes of the diamond bases. 2.3. Cuppable and Noncuppable Degrees
We say a degree a cups if there is a degree b such that a V b = 0', or equivalently, a has a complement with respect to V. In this subsection, we survey the results related to cupping.
2.3.1. Basic Results
As the natural dual notion of cappable degrees, one can define the notion of cuppable ones. We say that a c.e. degree a is cuppable, if there is an incomplete c.e. degree x such that a V x = O', and noncuppable, otherwise. The study of cupping and noncupping of degrees provides significant ingredients for undecidability and definability results for R. Sacks Splitting Theorem immediately implies the existence of cuppable degrees. However, the existence of noncuppable degree is entirely nontrivial and first proved by Cooper [18] and Yates(unpub1ished): Theorem 2.30. (Cooper and Yates) There exists a noncuppable c.e. degree a # 0 , namely, for a n y c.e. degree x, a V x = 0' if and only af x = 0'.
Let NCup denote the set of all noncuppable degrees. Clearly NCup is an ideal of R. This is the first known example of a nontrivial definable ideal of R.
2.3.2. Noncuppable Degrees and J u m p Classes Harrington [43] noted that the noncuppable degree in Theorem 2.30 can be chosen to be high and obtained the following extension of Theorem 2.30 (see Miller [75]):
282
Theorem 2.31. (Harrington) For every high c.e. degree h, there is a high c.e. degree a < h such that for any c.e. degree x, if h 5 aV x t h e n x s h . Consequently:
Corollary 2.4. (Harrington) Every high c.e. degree bounds a high noncuppable degree. Harrington also showed:
Theorem 2.32. (Harrington) Every degree is either cuppable o r cuppable. 2.3.3. Plus-Cupping Degrees
On the other hand, Harrington [44] (see Fejer and Soare [39]) also found a new class of degrees which behaves totally differently from the noncuppable ones. Theorem 2.33. (Plus-Cupping Theorem) There is a nonzero c.e. degree a such that for any c.e. degrees x , y , if 0 < x 5 a 5 y, t h e n there is a c.e. degree z < y such that x V z = y. We shall call the degree a in Theorem 2.33 a plus-cupping degree. Clearly, by Corollary 2.4, there is no high plus-cupping degree. Li [62] proved that the highness condition is necessary:
Theorem 2.34. There is a high2 plus-cupping degree. Using the properties of noncuppable and plus-cupping, Li [62] is able t o distinguish the classes HI and H, (n 2 2):
Corollary 2.5. For each natural number n > 1, TH(H1) # TH(H,), where TH(H,) is the elementary theory of upper semi-lattice H,. Proof. By Robinson’s [83] Interpolation Theorem, for any incomplete c.e. degree a, there is an incomplete high c.e. degree h > a. We say that a E H, is cuppable in H,, if there is an incomplete b E H, such that a V b = 0’, and noncuppable in H,, otherwise. By Theorem 2.34, for each
283
> 1, there
is a c.e. degree a E H, which bounds no degree noncuppable in H,. By Corollary 2.4, every high degree bounds a degree which is noncuppable in H1. 0 n
We can look for further relations between plus-cupping degrees and jump classes. For example, recently Li and Wang [68],showed that there exists a c.e. degree a # 0 such that for any c.e. degree x,if 0 < x 5 a, then x is lowz-cuppable, i.e., it is joined to 0' by a low2 c.e. degree.
2.3.4. Low,-Cuppable degrees Next we turn to the dual problem of Theorem 2.19: If a degree a is cuppable via some witness b, what is the lowest jump class which contains such a b? Li, Wu and Zhang [69] introduced the following notion: Definition 2.4. A c.e. degree a is called low,-cuppable, if there is a low, c.e. degree 1 such that a V 1 = 0'.
By Theorem 2.22, the set of noncappable degrees is equal to the set of all low-cuppable degrees. Li, Wu and Zhang [69] showed that we could go further beyond low-cuppable: Theorem 2.35. (Li, Wu and Zhang) The set of Zow-cuppable degrees is properly contained in the set of low2-cuppable degrees.
Theorem 2.35 also provides a natural definable relation F c R such that F n L1 = 0, and F n L2 # 8. Another interesting consequence is that from Theorem 2.35, we can derive the Harrington Cap and Cup Theorem: Theorem 2.36. (Harrington) There is a c.e. degree which is both cappable and cuppable.
Li [63] showed that the hierarchy of the cuppable degrees collapses eventually, in fact, at level low3: Theorem 2.37. (Li) If a degree a is cuppable then it is low3-cuppable.
However it is not known whether we can replace low3-cuppable by cuppable in Theorem 2.37. Question 2.3. Is every cuppable degree low2-cuppable?
1 0 ~ 2 -
284
2.3.5. Continuity of Cupping and Major Subdegree Problem Historically, Lachlan in 1967 suggested the notion of major subdegree, analogous to the notion of major subset in the context of the lattice of c.e. sets:
Definition 2.5. Given c.e. degrees a < b, we say that a is a major subdegree of b, if for every c.e. degree x, we have that b v x = 0’ holds if and only if a V x = 0‘ holds. Given that every noncomputable c.e. set B has a major subset (Lachlan [54]), it is natural to ask:
Major Subdegree Problem. Does every c.e. degree b # 0 or 0’ have a major subdegree? Attempts to solve the problem unexpectedly led to an appreciation of the importance of the notion of continuity for the c.e. degrees. We list a few results which lead to the solution of the major subdegree problem.
A theorem on continuity of cupping was provided by Ambos-Spies, LachIan and Soare [5]: Theorem 2.38. (Continuity of Cupping Theorem) For any c.e. degrees a, b, if 0 < a < 0’ and a v b = 0’, then there is a c.e. degree c < a such that c V b = 0’.
This theorem together with the Continuity of Capping Theorem (Theorem 2.24) gives a general continuity result. Theorem 2.39. (Harrington and Soare [46]) Let F(x,y) be a n open formula in the language {<, V, A, 0,O’) and a, b be distinct c.e. degrees such that F(a,b) holds in the upper semilattice R of c.e. degrees. T h e n there exist c.e. degrees ao,al,bo,bl such that (i) a0 < a < a l , bo < b < b l , and (ii) for any c.e. degrees x , y , if a0 F(x,y) holds.
< x < a1 and bo < y < bl, t h e n
Partial results directly related to the major subdegree theorem are obtained, notably by Cooper and Slaman [31], Seetapun [86] and Lachlan, Soare and Seetapun (see Seetapun [87]):
Theorem 2.40. (Cooper and Slaman)
285
(i) Every high c.e. degree h # 0‘ has a major subdegree. (ii) Every low c.e. degree 1 # 0 has a major subdegree.
Theorem 2.41. (Seetapun) Every nonzero low2 c.e. degree has a major subdegree. Theorem 2.42. (Lachlan, Seetapun and Soare) For any c.e. degree b # 0 , O‘, there exists a c.e. degree a 2 b such that for any c.e. degree x, i f 0‘ = b V x, then 0’ = a V x. This progress led to Cooper and Li’s full solution of the major subdegree problem. Cooper and Li [24] have shown:
Theorem 2.43. (Cooper and Li Continuity Theorem) Every c.e. degree b # 0,O’has a major subdegree, i.e., for any c.e. degree b # 0 ,O’, there is a c.e. degree a < b such that for any c.e. degree x, b V x = 0’ if and only i f a v x = 0’. An immediate consequence of the Cooper and Li Continuity Theorem (Theorem 2.43) is a continuity result for Harrington nonsplitting bases.
Corollary 2.6. (Cooper and Li) There is no minimal Harrington nonsplitting base. We end the topic with the open question below:
Question 2.4. Characterize the relations on R which are definable in R and which have maximal and for minimal elements. In particular, is there a relation on R which is definable in R and which has a nontrivial greatest or least element? 2.4. Definable Ideals 2.4.1. Definable Sets in R One of the central problems concerning R is to characterize the definable relations of R. It is closely linked to the rigidity problem of R,since the existence of certain definable relations may rule out certain automorphisms of R. Many problems on definability remain open. In this section, we give a survey of some partial results known so far.
286
We begin with the definable subsets. Using first order representation of standard models of arithmetic within R,Nies, Shore and Slaman [79] have shown: Theorem 2.44. For every n
> 0, H, and L,+1 are definable in R.
The recent work of Cholak, Downey and Walk [13] has provided another interesting example of definability, based on the definability of the relation “x is contiguous”, which can be characterized by the following theorem: Theorem 2.45. (i) (Downey and Lempp [35]) A c.e. degree a is contiguous if and only if it is locally distributive, i.e., if b V c = a > d, then there are bo, co such that bo < b, co < c and bo V co = d. (ii) (Ambos-Spies and Fejer [3] ) A c.e. degree a is contiguous if and only if it is not the top of an embedding of the lattice Ns into
R. This leads to the following definability of an anti-chain in R by Cholak, Downey and Walk [13]: Theorem 2.46. (i) There is a maximal contiguous degree, and furthermore, there is a c.e. degree a < 0’ such that there is a unique maximal contiguous degree > a. (ii) There is a definable infinite anti-chain of R. By Sacks Splitting Theorem, 0‘ is the join of two low degrees, hence L1 is not an ideal. Thus none of the jump classes forms an ideal.
2.4.2. Nies’ Results and Its Consequences
As discussed before, NCup and M are definable ideals in R; and by the two theorems of Harrington (Theorems 2.32,2.36), NCup is properly contained in M. For a long period of time, NCup and M are the only two known examples of definable ideals in R. The question was raised (again) by Shore in Boulders meeting [go]:
287
0
Are there other (naturally) definable ideals? Suppose that A is a definable subset. Is the ideal generated by A definable?
Andr6 Nies [77] made the breakthrough.
Theorem 2.47. (Nies) A n y definable subset of R generates a definable ideal of R. It should be noted that the proof was based on coding techniques, hence the resulting definition is not LLnaturaln. Thus there are many candidates for new definable ideals. However, if we are not careful, they may be some known ideals in disguise, as the following simple example shows. Let us call a degree a a diamond top, if a is the join of a minimal pair. Let DT' be the ideal generated by diamond tops. Then it is easy to see that DT' = M. Nies offered a new example of definable ideals by studying the nonbounding degrees. Let NB be the set of all nonbounding degrees. By a result of Ambos-Spies and Soare [6],NB is not closed under join (in fact the join of two nonbounding degrees can be high), hence not an ideal.
Theorem 2.48. (Nies) Let NB' be the ideal generated by NB. T h e n NB+ is a new definable ideal, more precisely,
(i) NB+ # Ad, hence NB' (ii) NB+ NCup.
cM.
The question whether or not NCup is contained in NB' was settled by the following result of Yang and Yu [98].
Theorem 2.49. (Yang and Yu) (i) NCup NB+. (ii) The ideal generated b y NCup and NB is n o t equal t o M. Consequently, the j o i n and intersection of NCup and NB+ are n e w definable ideals.
Let PC be the set of all plus-cupping degrees. It is interesting to compare the sets PC and NB. The two sets have many similarities, for example, both are properly contained in M and both are disjoint with the
288
high degrees. By a recent result of A. Li and D. Li [66], there exists a minimal pair a, b whose join is in PC, consequently PC # NB. Moreover, A. Li and Zhao [72] showed that PC is not closed under join by joining two plus-cupping degrees to a high one, hence PC is not an ideal.
Question 2.5. Does the ideal generated by PC equal to NB’? 2.4.3. Join Theorem in R and Its Consequences Although the study of the ideals generated by definable sets has its own merit, apparently it is too inefficient to get (infinitely many) new definable ideals. Perhaps we can make use of the definability of jump classes and produce a sequence of new definable ideals. We offer two such (failed) attempts below. Attempt one is to make use of the classes H,, and to generalize the notion of noncuppable degrees as follows. Let NCupH = {a : Vw(a V w E HI + w E HI}. Similarly, we can define NCupH, by replacing “H1”by “H,”.
It follows from Theorem 2.44 that for every natural number n, NCupH, is a definable ideal. The second attempt is to make use of the classes L, following the idea of deep degrees. Cholak, Groszek and Slaman [14] introduced the notion of almost deep degree.
Definition 2.6. A degree a is called an almost deep degree if V1 E Ll[(a v 1) E Ll].
In other words, joining with a preserves lowness. They showed:
Theorem 2.50. There is a nonzero almost deep degree. Notice that the almost deep degrees form an ideal. It is not known if it is definable. However, one could modify the notion to get definability as follows.
289
Definition 2.7. A degree a is called an almost n-deep degree if V l E L,[(a V 1) E L,].
In other words, joining with a preserves low,-ness. Notice that for n > 1, by Theorem 2.44 again, the almost n-deep degrees do form definable ideals. It was once believed that the method of proving Theorem 2.50 could be modified to produce a nonzero almost 2-deep degree. The notion of almost deep degree comes from “deep degree”. Historically, Bickford and Mills [12] defined a degree a to be deep if
Vw[(a v w)’ = w‘]. In other words, joining with a preserves the jump.
It can be generalized to n-deep degrees. We say that a degree a is n-deep if Vw[(av w)(,) = w‘”)]. In other words, joining with a preserves the n-th jump. Lempp and Slaman [59] showed that there is no nontrivial deep degree.
Theorem 2.51. (Lempp and Slaman) T h e only deep degree is 0. However, their argument did not settle whether or not there are nontrivial n-deep degrees. The attempts of using NCupH, and n-deep degrees do not work, as Jockusch, Li and Yang [51] have shown in the following version of the Join Theorem:
Theorem 2.52. (Jockusch, A. Li and Y. Yang) (Ma # 0)(3b)[b”= (a V b)‘ = O“]. Equivalently, (Va # 0)(3b E La)[(aVb) E HI]. ,
Corollary 2.7. (i) For each n _> 1, the ideal NCupH, is trivial. (ii) For each n >_ 1, the only n-deep degree is 0. (iii) For each n _> 2, the ideal of almost n-deep degrees is trivial. Theorem 2.52 has other interesting consequences. The original motivation came from the study of the join operator in c.e. degrees, in the same spirit as Posner and Robinson [go].
290
Theorem 2.53. (Posner and Robinson) In Turing degrees, for any 0 # a 5 0’, there exists b such that b’ = a V b = 0‘. However, by the existence of noncuppable degrees, we have no such join theorem in R. Thus, Theorem 2.52 is the optimal result one can hope for in R. Since there exists nonzero almost deep degree and there is no nonzero almost n-deep degree (for n > 1), we get the following elementary difference between L1 and L, ( n > 1):
Corollary 2.8. For each n tarily equivalent.
> 1, the structures L1 and L, are not elemen-
As we have seen, the complement of low-cuppable degrees is the cappable degrees (by Theorem 2.22), and the complement of lows-cuppable degrees is the noncuppable degrees (by Theorem 2.37). Both of them are ideals of R. We conjecture that the complement of lowz-cuppable degrees is also an ideal of R. (Of course, if the answer for Open Question 2.3 is positive, then the complement becomes the ideal NCup.) The overall picture of definable ideals is still unclear. There are many questions along the line of characterizing (at least, finding more) definable ideals of R. We list some of them below, the first one was attributed to Slaman.
Question 2.6. (i) Is there any nontrival principal ideal of R which is definable in R? (ii) Are there infinitely many definable ideals in R?.
2.5. Open Questions We end our discussion on c.e. degrees by some open questions. Many of them have been pointed out before, we select a few of them below.
Question 2.7. (i) Characterize the finite relations of R which are definable in R. (ii) Characterize all definable relations of R.
291
As we have seen, many theorems, for instance, Cooper's Minimal Pair Theorem (Theorems 2.15) and Robinson Splitting Theorem (Theorem 2.3), suggested the natural question of characterizing the jump classes via structural properties of R.
Question 2.8. (i) Is there any natural definable relation F of R such that F contains no high, c.e. degrees, but does contain high,+l c.e. degrees for any n > l ? (For n = 1, the answer is yes by existing results.) (ii) Is there any natural definable relation F of R such that F contains no low, c.e. degrees, but does contain low,+l c.e. degrees for any n > 2? (For n = 1,2, the answer is yes by existing results.) (iii) Is there a natural definable relation F c R of R which contains neither low,+l nor high, c.e. degrees for each n > O?
So far, one of the obstacles is the lack of methods. The powerful 0"'priority arguments normally produces sets in either H2 or L3. We could also ask: Question 2.9. Is there a natural problem which naturally requires a 0(4)priority argument? By Corollaries 2.8, 2.2 and 2.5, we can distinguish L1, LZ and HI from other jump classes by elementary properties. However we do not know how to distinguish other jump classes. We raise the question in a way that suggests the possibility that some jump classes might be elementarily equivalent.
Question 2.10. (i) Are there any m # n such that TH(H,) = TH(H,)? (ii) Are there any m # n such that TH(L,) = TH(L,)? The ultimate goal is to characterize jump classes by algebraic properties of R. We repeat the problem as stated in Shore [go]:
Question 2.11. Are any of the jump classes H, and L,+I naturally definable in R?
292
3. Difference Hierarchy The n-c.e. Turing degrees provide a natural generalization of the class of the computably enumerable degrees. One fascinating aspect is the comparison between the structures of the c.e. degrees and the n-c.e. degrees for n > 1 (and the Aa-degrees as well, which is beyond this paper). Recent years have seen much research and various similarities and differences have been found. We use V, to denote the class of all n-c.e. Turing degrees, for n > 1, and d.c.e degrees to denote elements in V2. 3.1. Cupping and Capping One of the first significant results concerning the structure of the n-c.e. degrees for n > 1 was provided by Arslanov [7].
Theorem 3.1. (Arslanov’s Cupping Theorem) Every nonzero c.e. degree can be joined t o 0‘ by a n incomplete d.c.e. degree. By the existence of noncuppable c.e. degrees (Theorem 2.30), this gives a difference between the elementary theories of R and V n for every n > 1. Notice that the difference is articulated by a Cs-sentence. Cooper, Lempp and Watson [23] extended Theorem 3.1:
Theorem 3.2. (Cooper, Lempp and Watson) For any noncomputable d.c.e. degree d and any high c.e. degree h > d, there is a low d.c.e. degree 1 such that h = d V 1. In contrast to Lachlan’s Nondiamond Theorem (Theorem 2.26), Downey [34] gave another difference, which is a C2-one:
Theorem 3.3. (Downey’s Diamond Theorem) T h e diamond lattice can be embedded into V2 preserving both 0 and 0‘. Jiang [50]showed that 0’ in Theorem 3.3 can be replaced by any high c.e. degree. The strongest result along this line is obtained by Li and Yi [71] which extends both Arslanov’s Cupping Theorem and Downey’s Diamond Theorem:
Theorem 3.4. (Li-Yi Cupping Theorem) There exist incomplete d.c.e. degrees ao,al such that f o r any c.e. degree x, i f x # 0, t h e n 0‘ = ai V x holds f o r some i 5 1.
293
3.2. Splitting and Density Sacks’ Splitting Theorem (Theorem 2.1) was among the first properties determined for R. The same cannot be said for d.c.e. degrees. The splitting property was generalized to the d.c.e. degrees in early the 1990’s by Cooper [20]:
Theorem 3.5. (Cooper) For any d.c.e. degree a # 0 , there are d.c.e. degrees b, c # 0 such that b < a, c < a and b V c = a. However this splitting theorem cannot be combined with cone avoidance, unlike the Sacks Splitting Theorem for the c.e. degrees. Cooper and Li [28], [29] have shown:
Theorem 3.6. (Cooper and Li) Given n > 1, there exists an n-c.e. degree a and a c.e. degree b such that 0 < b < a and such that for any nc.e. degrees x,y, i f x V y = a, then either b 5 x or b 5 y . A basic analysis of the proof of Theorem 3.6 establishes the following:
Theorem 3.7. (Cooper and Li [29]) For any n > 1, the elementary theory of the low3 c.e. degrees is diflerent from that of the lows n-c.e. degrees. Another basic and related topic is the density problem. For the c.e. degrees, density is established by Sacks [85] (Theorem 2.2). But the density problem for the n-c.e. degrees was much harder. Cooper, Lempp and Watson [23] proved a weak density theorem that for any n > 0, and any c.e. degrees c < a, there exists a properly n-c.e. degree b such that c < b < a. Cooper and Yi (unpublished, for n = 2) and Arslanov, LaForte and Slaman [lo] (for n > 2) proved another weak density theorem which says that for n > 1, n-c.e. degree b, and c.e. degree a, if a < b, then there exists a d.c.e. degree c with a < c < b. While Cooper, Harrington, Lachlan, Lempp and Soare [22] proved the nondensity of n-c.e. degrees for each n > 1:
Theorem 3.8. (Nondensity Theorem of n-c.e. Degrees) For every n > 1, there is a maximal n-c.e. degree in the n-c.e. degrees. The relation between the maximal degrees and the jump classes is not known, except the fact that it cannot be low by Arslanov, Cooper and Li [9] (Theorem 2.11).
294
Even though the structure of d.c.e. degrees is not dense, many weak density results, some of which are about set splitting instead of degree splitting, have been shown to hold. As an analogy t o Harrington-BickfordMills Theorem on low2-splitting (Theorem 2.9), Cooper [19] showed the following: Theorem 3.9. (Cooper) For each n > 0, each n-c.e. degrees a < 1, i f 1 is low, then there are n-c.e. degrees x,y such that a 5 x, y < 1 and 1 = x V y .
Considering splitting in the d.c.e. degrees, Ding and Qian (unpublished, see Ding, Lu and Qian [73]) and independently Ishmukhametov [48] showed: Theorem 3.10. For any c.e. sets B with the following properties:
(i) X O@ Xi ET A, and (ii) A $T X i @ B holds f o r each i
A, there exist d.c.e. sets X O ,Xi
5 1.
Furthermore, Cooper and Li [27] showed: Theorem 3.11. (Cooper-Li Splitting Theorem) For any d.c.e. set A, c.e. set B, if B
(1) X O@ Xi i~A, and (2) f o r any i 5 1, if A
Proof. By Theorem 3.11, there exist d.c.e. sets X 0 , X l such that either X 0 , X l satisfy the conditions of the theorem or there is a c.e. set I such
295
that A ET I @ B. In the latter case, by Theorem 3.10, there are d.c.e. sets YO, YI satisfying the conditions of Theorem 3.12. The degree theoretic statement of Theorem 3.12 is then:
Theorem 3.13. (Cooper and Li [27]) For any d.c.e. degree a, c.e. degree b, if b < a, then there are d.c.e. degrees xo,x1 such that b < x0,xl < a and xg v x1 = a. This offers us a candidate for defining R in D2.
Question 3.1. (i) Does the property in Theorem 3.13 define R in Ds?Namely, is it true that for any d.c.e. degree c , there is a d.c.e. degree d > c such that d does not properly split over c in V2? (ii) (Cooper and Yi [32]) Is R definable in Dn (for n > l)? From the discussion above, it seems that density and splitting always exist together.
Question 3.2. Is there a linearly ordered dense interval in Vz? 3.3. Isolation Pairs Cooper and Yi [32] and independently Ishmukhametov [47], showed that for any c.e. degree a, and d.c.e. degree d, with a < d, there is a d.c.e. degree c such that a < c < d. Also in [32], Cooper and Yi showed that c cannot be restricted to the c.e. degrees. The notion of an isolated d.c.e. degree was introduced by Cooper and Yi.
Definition 3.1. A d.c.e. degree d is isolated by a c.e. degree a, if a < d and any c.e. degree c below d is also below a. The pair a and d is called an isolation pair. We say that d is isolated if there is a c.e. degree a such that a isolates d, and nonisolated if there is no such c.e. degree,
Theorem 3.14. (Cooper and Yi) Both isolated and nonisolated degrees exist.
296
Notice that the existence of isolated degrees can also be obtained from Kaddah’s Infima Theorem [52] which says that each low c.e. degree is branching in the d.c.e. degrees. Thus any low nonbranching (in ‘R) c.e. degree a isolates both do and dl that have infimum a in V2. Since they were introduced, the notions of isolated and nonisolated degrees have been studied intensively. Their significance lies in both themselves and in their relationship of relative definability. It is now known that:
Theorem 3.15. (i) (Ding and Qian [33]; (LaForte [57] ) T h e isolated d.c.e. degrees are dense in the c.e. degrees. (ii) (Arslanov, Lempp and Shore [ll]) T h e nonisolated d.c.e. degrees are dense in the c.e. degrees. Cooper and Yi [32] asked whether each c.e. degree isolates some d.c.e. degree. Arslanov, Lempp and Shore [ll]provides a negative answer:
Theorem 3.16. (Arslanov, Lempp and Shore) There is a c.e. degree a isolating n o d.c.e. degree. We shall call the degree a in Theorem 3.16 a nonisolating degree. Arslanov, Lempp and Shore [ll]also showed that nonisolating degrees are downward dense in the c.e. degrees and are realized in each jump class. Recently, Wu and Ishmukhametov [49] obtain the optimal result concerning isolation pairs and jump classes:
Theorem 3.17. There is a n isolation pair a and d such that a is low and d is high. Wu [97] obtained other interesting results:
Theorem 3.18. There is a high nonisolated d.c.e. degree such that all c.e. degrees below it are bounded by a low d.c.e. degree. Wu [96] showed that there exist c.e. degrees a , c # 0, d.c.e. degree d such that a A c = 0, a V d = 0‘, c < d and c, d is an isolation pair. Hence ( 0 , a, d, 0 ’ ) is a proper diamond. By combining the isolation technique and the gap/co-gap method, Downey, Li and Wu [37] have strengthened the result and shown that in fact there are plenty of diamonds in 2)~:
297
Theorem 3.19. (Downey, Li and Wu) For any computably enumerable degree c # 0, if it is cappable in the computably enumerable degrees, then there exists a c.e. degree a and a d.c.e. degree d such that a < d, c v d = 0’, c A a = 0 and a, d is a isolation pair. Consequently, ( 0 , a,d, 0’} is a proper diamond in Vz. 3.4. Open Questions
The most well-known and longstanding question is the following, a positive solution to which is known Downey’s conjecture: Question 3.3. Are there m to Vn?
# n such that V, is elementarily equivalent
Another basic question was raised by Cooper, which generalize the Open Question 3.1: Question 3.4. At what levels of the difference hierarchy does Turing definability of given lower levels of the hierarchy occur? We would like to see more studies on jump classes in D,: Question 3.5. (i) Are the elementary theories of low, c.e. degrees and of low, nc.e. degrees different for each n > 1, and for m = 1 or 2? (ii) Are there any jump classes high,, low, n-c.e. degrees definable in Dn for n > l?
It is natural to speculate whether the accumulation of knowledge of the local degree structure contributes to a global understanding of the structure of Turing degrees. If we accept the thesis that doing fieldwork is necessary for a good understanding of the ecology of rain forests, then we would expect facts about local degree structures to play a major role in the solution of major global degree-theoretical problems in the future. References 1. Klaus Ambos-Spies. An extension of the nondiamond thoerem in classical and a-recursion theory. J. Symbolic Logic, 49:586-607, 1984.
298 2. Klaus Ambos-Spies. On pairs of recursively enumerable degrees. Trans. Amer. Math. SOC.,283:507-531, 1984. 3. Klaus Ambos-Spies and Peter A. Fejer. Embeddings of N5 and the contiguous degrees. Ann. Pure Appl. Logic, 112(2-3):151-188, 2001. 4. Klaus Ambos-Spies, Carl G. Jockusch, Jr., Richard A. Shore, and Robert I. Soare. An algebraic decomposition of the recursively enumerable degrees and the coincidence of several degree classes with the promptly simple degrees. Trans. Amer. Math. SOC.,281:109-128, 1984. 5. Klaus Ambos-Spies, Alistair H. Lachlan, and Robert I. Soare. The continuity of cupping to 0'. Ann. Pure Appl. Logic, 1993. 6. Klaus Ambos-Spies and Robert I. Soare. The recursively enumerable degrees have infinitely many one-types. Ann. Pure Appl. Logic, 44(1-2):l-23, 1989. Third Asian Conference on Mathematical Logic (Beijing, 1987). 7. Marat M. Arslanov. Structural properties of the degrees below 0'. Dokl. Nauk. SSSR, pages 270-273, 1985. 8. Marat M. Arslanov. Open questions about the n-c.e. degrees. In M. Lerman P. Cholak, S. Lempp and R. A. Shore, editors, Computability Theory and Its Applications: Current Trends and Open Problems, volume 257 of Contemporary Mathematics, pages 15-22, Providence RI, 2000. AMS. 9. Marat M. Arslanov, S. Barry Cooper, and Angsheng Li. There is no low maximal d.c.e. degree. Math. Log. Q., 46(3):409-416, 2000. 10. Marat M. Arslanov, Geoffrey L. LaForte, and Theodore A. Slaman. Relative enumerability in the difference hierarchy. J. Symbolic Logic, 63(2):411-420, 1998. 11. Marat M. Arslanov, Steffen Lempp, and Richard A. Shore. On isolating r.e. and isolated d-r.e. degrees. In Computability, enumerability, unsolvability, volume 224 of London Math. SOC.Lecture Note Ser., pages 61-80. Cambridge Univ. Press, Cambridge, 1996. 12. M. Bickford and C. F. Mills. Lowness properties of r.e. degrees. Unpublished preprint, 1982. 13. Peter Cholak, Rod Downey, and Stephen Walk. Maximal contiguous degrees. J. Symbolic Logic, 67(1):409-437, 2002. 14. Peter Cholak, Marcia Groszek, and Theodore A. Slaman. An almost deep degree. J. Symbolic Logic, 66(2):881-901, 2001. 15. Peter A. Cholak, Steffen Lempp, Manuel Lerman, and Richard A. Shore, editors. Computability Theory and Its Applications: Current Trends and Open Problems, volume 257 of Contemporary Mathematics, Providence, RI, 2000. American Mathematical Society. 16. S. Barry Cooper. Jump equivalence of the A! hyperhyperimmune sets. J. Symbolic Logic, 37:598-600, 1972. 17. S. Barry Cooper. Minimal pairs and high recursively enumerable degrees. J. Symbolic Logic, 39:655-660, 1974. 18. S. Barry Cooper. On a theorem of C. E. M. Yates. Handwritten notes, 1974. 19. S. Barry Cooper. Definability and global degree theory. In J. Oikkonen and J. Vaananen, editors, Logic Colloquium '90, volume 2 of Lecture Notes in Logic, pages 25-45, 1991.
299
20. S. Barry Cooper. A splitting theorem for the n-recursively enumerable degrees. Proc. Amer. Math. Soc., 115:461-472, 1992. 21. S. Barry Cooper. Local degree theory. In Handbook of Computability Theory, pages 121-153. North-Holland, Amsterdam, 1999. 22. S. Barry Cooper, Leo Harrington, Alistair H. Lachlan, Steffen Lempp, and Robert I. Soare. The d.r.e. degrees are not dense. Ann. Pure Appl. Logic, 55(2):125-151, 1991. 23. S . Barry Cooper, S. Lempp, and P. Watson. Weak density and cupping in the d-r.e. degrees. Israel J. Math., 67(2):137-152, 1989. 24. S. Barry Cooper and Angsheng Li. On Lachlan's major subdegree problem. to appear. 25. S. Barry Cooper and Angsheng Li. Splitting and nonsplitting, I: A low3 Harrington nonsplitting base. to appear. 26. S. Barry Cooper and Angsheng Li. Splitting and nonsplitting, 11: A low2 c.e. degree above which 0' is not splittable. to appear in JSL. 27. S. Barry Cooper and Angsheng Li. Turing definability in the Ershov hierarchy. to appear in Journal of London Mathematical Society. 28. S. Barry Cooper and Angsheng Li. Non-uniformity and generalised Sacks splitting. Acta Math. Sin. (Engl. Ser.), 18(2):327-334, 2002. 29. S. Barry Cooper and Angsheng Li. Splitting and cone avoidance in the d.c.e. degrees. Science in China (series A), 45(9):1135-1146, 2002. 30. S. Barry Cooper, Angsheng Li, and Xiaoding Yi. On the distribution of Lachlan nonsplitting bases. Arch. Math. Logic, 41(5):455-482, 2002. 31. S. Barry Cooper and Theodore A. Slaman. Major subdegrees and cup classes. Recursive Function Theory Newsletter, Item 371, Issue No. 37, 1988. 32. S. Barry Cooper and Xiaoding Yi. Isolated d.r.e. degrees. University of Leeds, Dept. of Pure Math., 1995. Preprint series, No.17. 33. Decheng Ding and Lei Qian. Isolated d.r.e. degrees axe dense in r.e. degree structure. Arch. Math. Logic, 36(1):1-10, 1996. 34. Rodney G. Downey. d-recursively enumerable degrees and the nondiamond theorem. Bull. London Math. Soc., 21:43-50, 1989. 35. Rodney G. Downey and StefTen Lempp. Contiguity and distributivity in the enumerable Turing degrees. J. Symbolic Logic, 62(4):1215-1240, 1997. 36. Rodney G. Downey, Steffen Lempp, and Richard A. Shore. Highness and bounding minimal pairs. Math. Logic Quart., 39(4):475-491, 1993. 37. Rodney G. Downey, Angsheng Li, and Guohua Wu. Complementing cappable degrees in the difference hierarchy. to appear. 38. Peter A. Fejer. Branching degrees above low degrees. Trans. Amer. Math. SOC.,273:157-180, 1982. 39. Peter A. Fejer and Robert I. Soare. The plus-cupping theorem for the recursively enumerable degrees. In Logic Year 19799-80: University of Connecticut, pages 49-62, 1981. 40. R. M. Friedberg. Two recursively enumerable sets of incomparable degrees of unsolvability (solution to post's problem, 1944). Proc. Nat. Acad. Sci. U.S.A., 43:236-238, 1957. 41. M. B. Giorgi. Continuity properties of degree structures. PhD thesis, Uni-
300 versity of Leeds, 2001. 42. Edward R. Griffor, editor. Handbook of Computability Theory, volume 140 of Studies in Logic and the Foundations of Mathematics. North-Holland Publishing Co., Amsterdam, 1999. 43. Leo A. Harrington. On Cooper’s proof of a theorem of Yates. notes, 1976. 44. Leo A. Harrington. Plus-cupping in the recursively enumerable degrees. notes, 1978. 45. Leo A. Harrington. Understanding Lachlan’s monster paper. notes, 1980. 46. Leo A. Harrington and Robert I. Soare. Games in recursion theory and continuity properties of capping degrees. In Proceedings of the workshop on set theory and the continuum. Mathematical Sciences Research Institute, Berkeley, 1989. 47. Shamil Ishmukhametov. D.r.e. sets, their degrees and index sets. PhD thesis, Novosibirsk, Russia, 1986. 48. Shamil Ishmukhametov. Splitting above r.e. degrees. Fundamental Mathematics and Mechanics, 1:139-143, 1996. 49. Shamil Ishmukhametov and Guohua Wu. Isolation and the high/low hierarchy. Arch. Math. Logic, 41(3):259-266, 2002. 50. Zhigen Jiang. Diamond lattice embedded into d.r.e. degrees. Sci. China Ser. A , 36(7):803-811, 1993. 51. Carl G. Jockusch, Jr., Angsheng Li, and Yue Yang. A join theorem for the computably enumerable degrees. t o appear. 52. D. Kaddah. Infima in the d.r.e. degrees. Ann. Pure Appl. Logic, 62(3):207263, 1993. 53. Alistair H. Lachlan. Lower bounds for pairs of recursively enumerable degrees. Proc. London Math. SOC.(3), 16:537-569, 1966. 54. Alistair H. Lachlan. On the lattice of recursively enumerable sets. Trans. Amer. Math. SOC.,13O:l-37, 1968. 55. Alistair H. Lachlan. A recursively enumerable degree which will not split over all lesser ones. Ann. Math. Logic, 9:307-365, 1975. 56. Alistair H. Lachlan. Bounding minimal pairs. J. Symbolic Logic, 44:626-642, 1979. 57. Geoffrey LaForte. The isolated d.r.e. degrees are dense in the r.e. degrees. Math. Logic Quart., 42(1):83-103, 1996. 58. Steffen Lempp. A high strongly noncappable degree. J. Symbolic Logic, 53:174-187, 1988. 59. Steffen Lempp and Theodore A. Slaman. A limit on relative genericity in the recursively enumerable sets. J. Symbolic Logic, 54:376-395, 1989. 60. Steven D. Leonhardi. Generalized nonsplitting in the recursively enumerable degrees. J. Symbolic Logic, 62(2):397-437, 1997. 61. Manuel Lerman. Embeddings into the computably enumerable degrees. In Computability theory and its applications (Boulder, CO, 1999), volume 257 of Contemp. Math., pages 191-205. Amer. Math. SOC.,Providence, RI, 2000. 62. Angsheng Li. Elementary differences among jump hierarchies. t o appear. 63. Angsheng Li. A hierarchy characterisation of cuppable degrees. t o appear. 64. Angsheng Li. Bounding cappable degrees. Arch. Math. Logic, 39(5):311-352,
301
2000. 65. Angsheng Li. Definable relations on the computably enumerable degrees. In S. B. Cooper and S. S. Goncharov, editors, Computability and Models. Kluwer Academic Publishers, 2002. 66. Angsheng Li and Dengfeng Li. A minimal pair join to a plus cupping degree. to appear. 67. Angsheng Li, Theodore A. Slaman, and Yue Yang. A nonlow;! c.e. degree which bounds no diamond base. to appear. 68. Angsheng Li and Yong Wang. A hierarch for the plus cupping Turing degrees. t o appear. 69. Angsheng Li, Guohua Wu, and Zaiyue Zhang. A hierarchy for cuppable degrees. Illinois J. Math., 44(3):619-632, 2000. 70. Angsheng Li and Dongping Yang. A high diamond theorem. Journal of Software, 11(1):23-39, 2000. 71. Angsheng Li and Xiaoding Yi. Cupping the recursively enumerable degrees by d.r.e. degrees. Proc. London Math. SOC.(3), 79(1):1-21, 1999. 72. Angsheng Li and Yicheng Zhao. The plus cupping Turing degrees do not form an ideal. t o appear. 73. Hong Lu, Decheng Ding, and Lei Qian. A splitting with infimum in the d-c.e. degrees. Math. Log. Q., 46(1):53-76, 2000. 74. W. Maass. Recursively enumerable generic sets. J. Symbolic Logic, 47:809823, 1982. 75. D. Miller. High recursively enumerable degrees and the anticupping property. In Logic Year 1979-80: University of Connecticut, pages 230-245, 1981. 76. A. A. Muchnik. On the unsolvability of the problem of reducibility in the theory of algorithms. Dokl. Akad. Nauk SSSR, N. S . 108:194-197, 1956. 77. Andre Nies. Parameter definability is the r.e. degrees. To appear. 78. AndrC Nies. Definability in the c.e. degrees: questions and results. In Computability Theory and Its Applications (Boulder, CO, 1999), volume 257 of Contemporary Mathematics, pages 207-213. Amer. Math. SOC.,Providence, RI, 2000. 79. Andre Nies, Richard A. Shore, and Theodore A. Slaman. Interpretability and definability in the recursively enumerable degrees. Proc. London Math. SOC.(3), 77(2):241-291, 1998. 80. David B. Posner and Robert W. Robinson. Degrees joining t o 0'. J. Symbolic Logic, 46(4) :714-722, 1981. 81. Emil L. Post. Recursively enumerable sets of positive integers and their decision problems. Bull. Amer. Math. SOC.,50:284-316, 1944. 82. R. W. Robinson. Interpolation and embedding in the recursively enumerable degrees. Ann. of Math., 93:285-314, 1971. 83. R. W. Robinson. Jump restricted interpolation in the recursively enumerable degrees. Ann. of Math., 93:586-596, 1971. 84. Gerald E. Sacks. On the degrees less than 0'. Ann. of Math., 77:211-231, 1963. 85. Gerald E. Sacks. The recursively enumerable degrees are dense. Ann. of
302
Math., 80:300-312, 1964. 86. David Seetapun. Contributions to Recursion Theory. PhD thesis, Trinity College, Cambridge, 1991. 87. David Seetapun. Defeating red. notes, 1992. 88. Joseph R. Shoenfield. Applications of model theory to degrees of unsolvability. In J. W. Addison, L. Henkin, and A. Tarski, editors, The Theory of Models, Proceedings of the 1963 International Symposium at Berkeley, Studies in Logic and the Foundations of Mathematics, pages 359-363, Amsterdam, 1965. North-Holland Publishing Co. 89. Richard A. Shore. The recursively enumerable degrees. In Handbook of Computability Theory, volume 140 of Stud. Logic Found. Math., pages 169-197. Nort h-Holland, Amsterdam, 1999. 90. Richard A. Shore. Natural definability in degree structures. In Computability Theory and Its Applications (Boulder, CO, 1999), volume 257 of Contemp. Math., pages 255-271. Amer. Math. SOC.,Providence, RI, 2000. 91. Richard A. Shore and Theodore A. Slaman. Working below a low2 recursively enumerably degree. Arch. Math. Logic, 29(3):201-211, 1990. 92. Richard A. Shore and Theodore A. Slaman. Working below a high recursively enumerable degree. J. Symbolic Logic, 58(3):824-859, 1993. 93. Richard A. Shore and Theodore A. Slaman. Defining the Turing jump. Math. Res. Lett., 6(5-6):711-722, 1999. 94. Theodore A. Slaman. The global structure of the Turing degrees. In Handbook of Computability Theory, volume 140 of Stud. Logic Found. Math., pages 155-168. North-Holland, Amsterdam, 1999. 95. Robert I. Soare. Automorphisms of the lattice of recursively enumerable sets. Bull. Amer. Math. SOC.,80:53-58, 1974. 96. Guohua Wu. Isolation and lattice embeddings. J. Symbolic Logic, 67(3):1055-1064, 2002. 97. Guohua Wu. Structural Properties of D.C.E. Degrees and Presentations of C.E. Reals. PhD thesis, Victoria University of Wellington, 2002, 98. Yue Yang and Liang Yu. On the definable ideal generated by nonbounding c.e.degrees. to appear. 99. C. E. M. Yates. A minimal pair of recursively enumerable degrees. J. Symbolic Logic, 31:159-168, 1966. 100. C. E. M. Yates. On the degrees of index sets. Trans. Amer. Math. SOC., 121:309-328, 1966.
303
A LIMIT STAGE CONSTRUCTION FOR ITERATING SEMIPROPER PREORDERS
TADATOSHI MIYAMOTO Department of Mathematics Nanzan University 18, Yamazato-cho, Showa-ku Nagoya, 4668673 Japan e-mail: miyamoto@nanzan-u. a c . j p
50. Introduction
We have two classes of preorders, i.e., proper and semiproper preorders. Semiproper preorders are more comprehensive than proper ones and sometimes coincide with the class of preorders which preserve the stationary subsets of w1. The countable support iterations of proper preorders are proper themselves. For the purpose of iterating semiproper preorders, the revised countable support iterations have been devised. These preorders are known to have corresponding forcing axioms. We may find the latest on these in [3]. In the meantime, people have sought simpler ways to iterate semiproper ones. For example, [1],[2],[4] and others. In this note, we consider a theory on iterated forcing and preservations of semiproper preorders. This is a combination of p. 539 in [3]and [4].More specifically, this note is organized as follows; 31. We formulate objects called general iterations. 52. We study limit stages in general and define iterations. 53. We consider an operation on limit stages and formulate our proposed limit, called simple. This took an impetus from p. 539 of [3]. 54. This part is the same as [4]with the new limit. Although we do not presuppose anything from [4],it gets considerably simpler. 55. We describe recursive constructions at limit stages of simple iterations. 56. We quickly review definitions relevant to semiproper and prove
304
our preservation theorem. It also contains an observation on the chain conditions. This obervation has its relevance to a question on p. 68 of [5].
Preliminary: We say (P,5,l)is a separative preorder iff the binary relation 5 on P is reflexive and transitive with a greatest element 1and for any p, q E P, p 5 q iff p Ikp “q E G”, where G denotes the canonical P-name for the generic filters over the ground model V. Let D be a subset of P and p E P. We say D is predense below p iff p IFpL‘Dn G # 0’’iff for any q 5 p there is d E D s.t. q and d are compatible in P. Notational remarks follow. Preorders are frequently specified by just writing P. For p , q E P, we write p q to express q 5 p 5 q. Namely they are equivalent conditions. But since we work with preorders, there is no “cp” to need to take the equivalence classes. For a formula cp, we write abbreviate 1Ikp“cp”.For a sequence s, Z(s) denotes the length of s which is the same as the domain of s and is an ordinal. If a 5 Z(s), then s r a denotes the initial segment ( s ( i ) I i < a ) . If a 5 P 5 Z(s), then s r [ a , p ) denotes ( s ( i ) I a 5 i < P). If a is a sequence with Z(a) 5 P 5 Z(s), then a^sr[Z(a),/?) denotes a U sr[Z(a),P) = ( a ( i ) I i < Z(a)) u ( s ( i ) I Z(a) 5 i < P ) = {(i, .(ill I i < q a ) ) u ((4s ( i ) ) I l ( a ) 5 i < P } . 1. The General Iterations Let us begin by specifying the basic objects of our concern.
sa,
1.1 Definition. Let I = ((Pa, la) I a < u ) be a sequence of separative preorders s.t. for any a < u , Pa is a set of some sequences of length a. We say I is a general iteration, if for any a 5 /3 < u, the following hold. 0 0
lpra = 1, and if p E Pp , then p r a E Pa. If p E Pp, a E P, and a 5, pra, then a-pr[a,p) a-pr[a, PI P.
sP
E Pp and
sp
~ f p q, then Pra
qra and P
Pra-qr[a,p).
Notice that the general iterations include products, finite and countable support iterations. The following is immediate. 1.2 Proposition. Let I be a general iteration and have 0
If. E P,, then x ^ l a [ [ a , p )E Pa.
Q
5 P < u. T h e n we
305
For the sake of clarity, we state the following.
1.3 Corollary. Let I = (P, I a < u ) be a general iteration and a 5 p < u. Then the map from P, to Pp defined b y x e x n l is a complete embedding and the map from Pp to P, defined b y p e p[a is a projection. We consider quotients in order to manage the generic filters. 1.4 Definition. Let I = (Pa I a < u ) be a general iteration, a 5 p < u and G, be a Pa-generic filter over the ground model V . Then in the l a p ) as follows. generic extension V[G,], we define the quotient (Pap, pap= {p~-[a,p) I P E pp andpi-a E G,}, l a p= lpi-[a,P)and ~ [ [ Q , Psap ) p[[a,p) iff there is u E G, s.t. u n q [ [ a , p ) 50 unp[[a,P) iff there are u1 and u2 in G, s.t. u r q [ [ a , p ) I p u T p [ [ a , p ) . Then we may show (Pap,l a p , l a p ) is a separative preorder.
sap,
The general iterations enjoy the following one-to-one and onto correspondence between the generics.
1.5 Proposition. Let I = (Pa I a Then we have
< u ) be a general iteration and a 5 p <
u.
If Gp is Pp-generic over V , then {p[a I p E Go}, denoted b y Gp[a, is Pa-generic over V and {p[[a,p) I p E Go}, denoted b y G p [[a, p), is Pap-generic over V[Gp[ a ] . (one-to-one) If G, is Pa-generic over V and H is Pap-generic over V[G,], then { q E Po I q[a E G, and q [ [ a , p ) E H } , denoted by G, * H , is Pp-generic over V . And both (G, * H ) [a = G, and (G, * H ) [[a,p) = H hold. (onto) If Gp is a Pp-generic filter over V and p E Po, then p E Go m r a E (car(.) andpr[a,p)E ~ ~ a r [ a , ~ ) . 1.6 Note. Given p,q E Pp, to show p
q, we do the following.
Fix any Pp-generic filter Gp over V with p E G p . Find a < p, the value of a may depend on G p , so that q[a E Gp [a and qr[a, P ) E G~ [[a,PI.
306 0
Conclude that q E Gp. Since G p is arbitrary and Pp is assumed separative, we may conclude p 50 q.
2. The Limit Stages
We investigate basic facts on the limit stages (by this we mean P i with limit i) of general iterations and define what is called an iteration by putting the weakest restriction on the order relations at the limit stages. We begin with the following necessities on the limits of general iterations for clarity. 2.1 Proposition. Let J = (P, 1 a
5 u ) be any general iteration with the limit ordinal u. The limit stage P, satisfies the following. (The universe of the order) {pnl,[[a,u) I p E P, and a < u } G P, S {x 1 x is a sequence of length u s.t. for all a < u, xra E P,}.
<,
q, then for all a
0
(The order relation) If p
0
(The greatest element) For all a
< u,
< u, p [ a 5, q[a.
1,ra = 1,.
We observe the connection between the order relation and the generic filters at the limit stages of general iterations. 2.2 Proposition. Let J = (Pa 1 a 5 v) be any general iteration with the limit ordinal u . The following are equivalent, where G , denotes the canonical P,-name for the generic filters.
For any x E P,, I/-, “a: E G, iff V a < v xra E G,[a”. For any x, y E P,, y 5, x iff V a < u y [ a 5, x [ a . Accordingly, we make the following two definitions. 2.3 Definition. Let I = (P, I a < v) be any general iteration with the limit ordinal u. We define l,, I*, I* and < I . as follows.
0 0
1, = U { l a I a < u } . I , = {pnlv [[a,u ) I p E P, and a < v}. I* = {x I x is a sequence of length u s.t. for all a
< u, sra E Pa}.
307 0
For z , y E I * , z
51.y , if for all a < u, z [ a 5 ,
y[a.
It is routine to check that the following are general iterations. 0 0
I - ( ( I * , < I * [ I +l, u ) )( I , is called the direct limit of I ) . I n ( ( I * , < I * , 1,)) (I* is called the inverse limit of I ) .
2.4 Definition. A general iteration I = (P, I a if 0
For any limit pra qra.
p < u and any p , q
< u ) is called an iteration,
E Po, p
< p,
sa
For the sake of clarity, we record the following. 2.5 Proposition. Let I = (Pp I ordinal L,3 < u, the following hold.
p <
u)
be an iteration. For any limit
c
(UP)* G PO (UP)* and
E
Go iff for
Here is a necessary and sufficient condition to form a limit stage of an iteration. Since we observed the inverse limit is indeed a limit, the verification of this proposition takes nothing. 2.6 Proposition. Let I = (P, I a < u ) be any iteration with the limit ordinal u . For a n y X C I* with 1, E X , the following are equivalent. 0 0
I n ( ( X , < I * [X, l u ) )is an iteration. ( X is called a limit of I ) For any z E X , a < v and u E P,, if u 5, z [ a , then u n z [ [ a ,u ) E
X. Therefore we look for X I* with the set theoretic closure property as above to define a limit. Since any limit is necessarily a suborder of the inverse limit I* of I , we may omit mentioning the order relations on the limits. The direct limit I , is the &-least closure with 1, in it. More generally, we observe the following.
2.7 Proposition. Let I = (P, I Q < v) be any iteration with the limit ordinal u. For any z E I * , let us define (.) = {u-x[[a,u ) I a < u,u E P, and u 5 , z [a}U I,. Then we have
308
0
0
0
(The single element closure) -I n ( ( z ) ) is an iteration and x is in the limit. In particular, (1,) = I*. (The closure) For any X C I * , let us define = 1 zE X}. Then I n ( x ) is an iteration and X is contained in the limit. (The meet and join) If both I n ( X ) and I - ( Y ) are iterations, then so are I n ( X n Y ) and I n ( X U Y ) .
x u{(.>
This means that the set of possible limits is a lattice under the inclusion whose minimum is I* and whose maximum is I*. Now we may ask which ones with what properties are important and when ? We close this section with the following on the dense subsets and the intermediate stages.
2.8 Lemma. Let I = (P, I a < u ) be any iteration. Suppose i 5 a 5 j < u, G , is a Pa-generic filter ouer V and set Gi = G,ri as usual. If D E V [ G J is a dense subset of Pij, then for any x E Pj with x [ a E G,, there is y 5 z s.t. y [ a E G, and y[[i,j) E D . Proof. The case a = j suffices. The rest is standard.
0
3. The Dynamical Stages, Simple Limit and Iterations
The following is a modified version of p. 539 of [3]. 3.1 Definition. Let J = (Pa 1 a 5 u> be any iteration with the limit ordinal u . Any P,-name ci is called a P,-dynamical stage ( or simply stage), if the following two hold. 0
It-p""&
5 u".
If p I ~ p v " c i= omit - 's)
p , then p [ J n l , [ [ J , u ) (kpv"d!= p .
(We usually
The idea is we want ci < u in relevant cases and use the value as a stage of the iteration below u. We also demand ci = u in irrelevant cases. We do not just set ci = 0 in irrelevant cases. We present typical dynamical stages.
3.2 Proposition. Let J = (Pa I a 5 u ) be any iteration with the limit ordinal u. The following are typical. 0
(Checks) For any Q
5 u, the P,-name fi is a P,-dynamical stage.
309
0
(Max) If both d! and ,b are P,-dynamical stages, then so is any P,-name for the maximum (in Vpu) of these two ordinals.
Dynamical stages enjoy the following properties. 3.3 Proposition. Let J = (Pa I a 5 u ) be any iteration with the limit ordinal u . Then for any P,-dynamical stage d! and 7 5 u, we have independences as follows.
wdr,
itpv vcp(&,
If P II-~,, f j ) ’7, then PrV-1, r[q, f j ) 77, where “ q ( d ! , i j ) ” stands for any of “d! < f j ” , % = f j ” , ‘Yj < &”, “d! < q”7 or ‘Yj _< d!”. We consider conditions with associated dynamical stages as follows. The comments in the parentheses are easy to verify.
3.4 Definition. Let J = (Pa 1 a 5 u ) be any iteration with the limit ordinal v. For any x E P, and P,-dynamical stages (8, I n < w ) , we say (6, 1 n < w ) is a sequence of P,-dynamical stages for x, if the following three hold. 0 0
For any n < w , lkpu“8, 5 &+I 5 u ” . (increasing) For any n < w , x lkpv“h, < u” . ( x may not decide the values of &, but knows that they are all strictly less than u. This is equivalent then &, < Y”.) to lkp,, “if x [ & E G,[in, I~p,,“If8 = SUp{Jn I n < w } ( < u ) and x [ 8 E G,[8, then x E G,”. ( x [ b knows that the rest of x is equivalent to l,[[&,u)in 4 , . Namely, this is equivalent to ( k p ,“if 8 =sup{& 1 n < w } and x [ &E G, 18, then XI[&, u ) G 1, “8, u)”.)
In the above, we may also say that x has (a sequence of) P,-dynamical stages (8, I n < w ) . We say 6, is a lc-th P,-dynamical stage (or simply k-th stage) for x. It is clear that for any < u , (( 1 n < w ) is a sequence of P,-dynamical stages for 1,. In general, x E P, may or may not have any sequence of P,-dynamical stages.
<
We now observe some of the important properties of dynamical stages.
3.5 Lemma. (Hooking) Let J = (Pa I a 5 u ) be any iteration with the limit ordinal u. Suppose y 5 , x, (inI n < w ) and (rj, I n < w ) are P,-dynamical stages for x and y , respectively. Then we have P,-dynamical stages ( ~ 1kn < w ) for y s.t. for any n < w , Ikp,, ‘$,+I 5 rjk and rjn 5 rjk ”.
310
Proof. This is easy. Just let
ik
= the maximum of
(An+l,in}
in V p u . 0
We are free to move stages forward and choose a O-th stage. This is a specific case of Hooking above. 3.6 Corollary. Let J = (Pa 1 Q
5 u ) be any iteration with the limit ordinal
u. The following hold. 0
0
If x E P, has P,-dynamical stages (8, I n < w ) and E < u, then there are P,-dynamical stages (in1 n < w ) for x s.t. for any n < w , x I ~ P ,‘Gn, , 5 I +n ”If x E P, has P,-dynamical stages and < v, then we may prepare P,-dynamical stages (Ak I IC < w ) for x s.t. ~t-p,, 4 0 = E”.
<
We consider an operation on the limit stages. 3.7 Definition. Let J = (Pa1 Q 5 v) be any iteration with the limit ordinal v. The suborder D(P,) = {x E P,, 1 x has P,-dynamical stages} of P, is called the dynamical kernel of P,.
Sometimes nothing new happens.
3.8 Proposition. Let J = (Pa I ordinal u. The following hold. 0
0
Q
5
u)
be any iteration with the limit
If P, = ( J [ u ) * (i.e., P, is the direct limit of Jru), then D(P,) = p,. Ifcflv) = w , then D(P,) = P,.
The dynamical kernel provides us a limit and the operation D enjoys that D2 = D. Namely, P,-dynamical stages are turned D(P,)-dynamical stages. The following are both crutial. 3.9 Lemma. Let J = (Pa I Q 5 u) be any iteration with the limit ordinal u. And we write I = J [ u for short. (1) Then I n ( ( D ( P u ) 5 , , [D(P,), 1,)) is again an iteration. Namely, we have the closure as follows. 0 0
1, E D(P,,).
I f x E D(P,), a u ^ x [ [ a ,u ) E D(P,).
<
u, u E
Pa and u La
X[Q,
then
31 1
(2) And D(D(P,)) = D(P,) holds. Namely, we have 0
For any x E D ( P V ) ,there is a sequence of D(P,)-dynamical stages for x. (In general, given limits X, Y of I with X C Y and x E X, Y-dynamical stages for x naturally give rise to X-dynamical stages for x.)
Proof. Since we are free to choose a 0-th stage, it is easy to observe the closure properties so that D(P,) is indeed a limit. We provide some details to the latter half. Let X and Y be any limits of I with X C Y . Namely, both I - ( X ) and I - ( Y ) are iterations. In particular, both X and Y are suborders of I*. Suppose x E X has Y dynamical stages (&, I n < w ) . We may define a sequence (8, I n < w ) of X-names as follows, where G x denotes the X-generic filter over V .
i,
=
{ <,v,
<
if < v and a-1 otherwise.
11-y "&,
=
for some a E G X I<.
We claim this (8, I n < w ) is a well-defined sequence of X-dynamical stages for x. We only verify two conditions.
x I k X " 6 , < v".
<
Proof. Suppose x1 5 x in X . We want to find 2 2 5 x1 in X and < u s.t. 2 2 IFxLL6, = t". Since both x and x1 are in X and X C Y , these belong
to Y.By assumption, x IkyL'&,< v". So we have x1 I ~ Y " & , < u " . Hence there is (a,<) s.t. < u , a E Pc with a 5 XI[[ and a-1 Iky"&,= 5". Let x2 = anxl[[<, v). Since x1 E X, we have x2 E X . This 2 2 works. 0
<
0
Ikx "If h = sup{& I n < w } and x[8 E G X [&,then x E G x " .
Proof. Suppose x1 E X, S < v, x1 1kx"d-= sup{& I n < w } = 6" and x1rS 5 x[S. We show x1 5 x and this suffices. To this end, we first observe x1 [S-1 IFy"sup{&, I n < w } = 6" as follows. Let G y be any Y-generic filter over V with x1[6-1 E G y . We would like to calculate SUP{(&,)G~ I n < w } . TOdo this, we fix an X-generic filter G x over V s.t. 21 E G x and G x [S = G y [S. This is possible, since both I - ( X ) and I - ( Y ) are iterations. For each n < w , let = ( 8 , ) ~ ~in V [ G x ] . Since 5 6 < u , there is a, E Gxr<, s.t. a;l IkyLL&, = 5,". Notice that a z l E G y by the way we fixed G x . So ( & ) G ~ = <., Therefore (<, I n < w ) = ( ( & , ) G ~ I n < w ) E V [ G y ]and SUP{(&,)G~I n < w } = 6.
<,
<,
312
Since (& I n < w ) is a sequence of Y-dynamical stages for x and we have seen x1 [S-1 I~yi‘sup(ciYn I n < w } = 6 and x1 [S E GyrS”, we get x1 [S-1 Iky“x E Gy”. Hence x1rS-1 5 x. In particular, x1 5 x holds. 0 Let us close this section by defining the limit and iterations we are interested in. 3.10 Definition. Let I = (Pa I a < v) be any iteration with the limit ordinal v. The simple limit of I is the suborder D ( I * )of I* (i.e., ( D ( I * ) 51. , [ D ( I * )lv), , where D ( I * ) denotes the dynamical kernel of the inverse limit I*of I ) . 3.11 Definition. An iteration I = (Pa I a if the following two hold.
< v) is called a simple iteration,
+
(The fullness at the successors) For any a with a 1 < v, if p l k p , “ ~E P,+l with ~ [ Ea G:,”, then there is q E P,+l s.t. q[a = p and p I ~ P , “qr[a,a 1) ~ [ [ aa , 1 ) in P:,:,+l”. For any limit a < v , Pa is the simple limit of Ira.
+
0
+
The fullness at the successors differentiates the (simple) iterations from (mere) products. The usual recursive construction would satisfy this requirement. And we just recursively take the simple limit at every limit stage to construct our iteration. We sum up as follows. 3.12 Proposition. Let I = (Pa I a < v) be any simple iteration. For any limit ordinal j < v, the following hold.
0
0
( I [ j ) *C Pj 5 ( I [ j ) * For any x E ( I [ j ) * x, E Pj i$x has a sequence of (I[j)*-dynamical stages. And so if c f ( j ) = w , then Pj = ( I [ j ) * . (This is the same as countable support iterations. However at j with cf(j) 2 w1, we may get bigger Pj than the direct limit.) For a n y x , y E P j , x Lj y i f fVcu < j z[cu 5, y r a . For any i < j and x E Pj, x has Pj-dynamical stages (An I n < w ) s.t. Ikpj ‘80 = i”.
4. Nested Antichains and Fusion Structures
We would like t o introduce nested antichains. We motivate this object in relation to an analysis of dynamical stages as follows. Let J = (Pa 1 a 5 v)
313
be any (simple or not) iteration with the limit ordinal v. Suppose we have given a sequence (& I k < w ) of Pu-dynamical stages for some z E P,. We want to analyze this situation. The following construction would come up with a single dynamical stage. And this construction will be recursively nested, since we have countably many dynamical stages. 4.1 Proposition. Let J = (Pa I a 5 v) be any iteration with the limit ordinal v. If x E Pu, is a P,-dynamical stage , J < v and a E Pt s.t. a 5 x[J and a - z [ [ J , v ) lkpv 5 < v ” (and so we m a y replace a-z[[J,v) by a-l,[[<,v>), then there is B g U{Pa I a < v} s.t.
b
a a a
‘x b
For b E B , J = l ( a ) 5 l ( b ) < v, b 5 u ^ x [ [ J , l ( b ) ) and b-11t-p” = l(b)”. { b [ l ( a )I b E B } is a maximal antichain below a in Pl(a). For any b, b’ E B , b[l(a)= b‘[l(a) implies b = b’.
q
It is conceivable that B = { a } . And when B has more than one elements, the lengths l ( b ) may vary with the b’s. So a gets partitioned into B[J and each element in B [ J end-extented to a unique b. With this motivation in mind, we make the following. (This and the next defined in [4].) 4.2 Definition. Let I = (Pa I a < v) be any iteration with the limit ordinal v. A nested antichain in I is a triple (T, (T, I n < w ) , ( suc; I n < w ) ) s.t. a
0
0
T = U{Tn I n < w } . To = {ao} for some a0 E U{Pa I a < v}. For any n < w , T, g U{Pa I a < v} and suc;
: T, + P(T,+l). For any a E Tn and b E suc$(a), l ( a ) 5 Z(b) and b[l(a) 5 a. For any a E T,, {b[l(a)I b E suc;(a)} is a maximal antichain below a in Pl(,). For any a E T, and b, b‘ E SUC$(~), if brZ(a) = b’[Z(u),then b = b’. Tn+1 = U{SUC$(~) I a E Tn}.
We talk about nested antichains by just writing T for brevity. Now we move on to consider a binary relation between the set of nested antichains T and the inverse limit I*. The T’s in the domain of this relation specify equivalent elements in I* via this relation. Hence the T’s in the domain represents a set of conditions in I * . It is known that this set of conditions forms a limit stage of I and is called the nice limit in [4].
314
4.3 Definition. Let I = (Pa I a < u ) be any iteration with the limit ordinal u. For any nested antichain T in I , x E I* is called (T,I)-nice, if the following hold. 0
0 0 0
x [ l ( a o ) ao, where {ao} = TO. For any a E T,, a 5 xrl(a). For any a E T, and b E sucF(a), b = b[l(a)-x[[l(a),Z(b)). For any a < v and w Pa with w 5 x [ a , if w I t p a “ there is a sequence (a, I n < w ) (this is called a generic cofinal path through T in v[G:,]) s.t. a. E T ~for, any n < w , a, E T,, a,+l E s u c ~ ( a , ) , /(a,) 5 a and a, E Ga[l(a,)”, then w - z [ [ a , u ) = w - l u [ [ a , u ) .
Nested antichains with associated nice sequences can be regarded as conditions as follows. 4.4 Proposition. Let I = (Pa I a
< u ) be any iteration with the limit ordinal u . Suppose T is a nested antichain in I and x E I* is (T,I)-nice. If x E X and I n ( X ) is any iteration (i.e., X is any limit stage of I ) , then the following holds, where GX denotes the X-generic filter over V . 0
I ~ ‘xx
E G x i f f there is a generic cofinal path through T in
V[GX]”. Proof. Suppose x E G x . We construct a generic cofinal path by recursion. Since x rl(a0) ao, where ao E TO,we have a0 E G x [l(ao). Suppose we have gotten a, E T, n (Gx[l(u,)). Then there is a,+l E sucF(a,) s.t. an+1 ri(an) E Gxri(a,). B U ~a,+l a,+l [l(an)-xr[l(a,),l(a,+l)) and ~ ( a ~ + his ~ ) . xr[i(an),~ ( a , + ~ )E) G~ r[l(an),G , + ~ ) ) . so a,+l E completes the construction. Conversely, suppose we have a generic cofinal path (a, 1 n < w ) through T in V [ G x ] .Since a, 5 x [ l ( a n ) ,we have x [ l ( a n ) E Gx[l(a,). Set a = sup{l(a,) I n < w } . Then x [ a E Gxra. Now if a = u, then we are done. So suppose a < v. We note that (a, I n < w ) E V [ G x r a ] .This is because we may define the sequence via G x [a. So we may take w E G x [a s.t. w 1t-p- “there is a generic cofinal path through T in V[Ga]”.Hence wnxC[a, v) w n l u [ [ a , u )and so x E G x . 0
=
cx
Although it is more tedious than above, we may prove the following equivalence along the same line. 4.5 Proposition. Let I = (Pa I a
<
u)
be any iteration with the limit
315
ordinal Y and T be a nested antichain in I . For x E I * , the following are equivalent. 0
x is ( T ,I)-nice. For any a < v, lkp? “xra E G, i f f (either there is a E T s.t. Q < l ( a ) and ara E G, or there is a generic cofinal path through T in V[G,])”.
As we see below nice sequences belong to dynamical kernels. 4.6 Proposition. Let I = (P, 1 a! < u ) be any iteration with the limit ordinal u. If x E I* is ( T ,I)-nice for a nested antichain T in I , then x E D( I *) . (i.e., x belongs to the simple limit of I.)
Proof. We must exhibit I*-dynamical stages (b, I n < w) for x. For each n < w,we define an I*-name b, as follows, where GI* denotes the I*-generic filter over V .
&+I =
if a E T,, b E suc$(a), b[l(a) E otherwise.
We claim this (b, I n and left.
GI* [Z(a) for some a and b.
< w)is well-defined and works. Details are routine 0
Now we move on to consider a type of basic construction at the limit stages of iterations. We would like to introduce what we call fusion structures. For the sake of motivation, we first mention a typical diagonal construction.
4.7 Proposition. Let J = (P, I Q 5 w) be a simple iteration. (So J is a usual countable support iteration of length w.)If (P, I n < w) is a sequence of elements of P, s.t. f o r all n < w,p,+l 5 p , and pn+l [(n 1) = p n [ ( n l), then there is p E P, s.t. for all n < w,p 5 p , hold. In fact, we may coastructp = (pn(n)I n < w) so that f o r all n < w,p [ ( n 1) = p , [(n 1) get satisfied.
+
+
We generalize the above as follows. (This defined in [4].)
+
+
316
4.8 Definition. Let J = (PaI a 5 v) be any iteration with the limit ordinal v and T be any nested antichain in Jrv. A structure F = ((a,n)I+ ( X ( ~ I ~(df’”) ) , I k < w))1 a E T,,n < w)is called a fusion structure in J , if the following three hold. 0 0
0
(6f’”) I k < w)is a sequence of P,-dynamical stages for x ( a , n )E P,. a 5 rl(a) and a-1 lkp,,“&g’n) = Z(a)”. (a-1 decides the ~
(
~
1
~
)
value of the 0-th stage for x ( ~ > ~ ) ) For any b E sucF(a) and k < w, and Ikp,, 5 < &f“n+l)”. ( With the Boolean value 1, the k-th stage for x(b,n+l) is ahead of and bounds the (k 1)-st stage for x ( ~ , ~ ) . )
“&hy)
+
A condition y E P, is called a fusion of the fusion structure F,if y is (T,J [v)-nice. 4.9 Lemma. (Fusion) Let J = (Pa I a 5 v) be any iteration with the (8f’”’ I k < w))I a E T,,n < w) limit ordinal v. If F = ((a,n)c) is a fusion structure in J with a fusion y E P,, then y Ikp,, “there is a generic cofinal path (a, I n < w) through T in V[G,] s.t. for each n < w, (
X(GL>4E
G,
~
(
~
1
~
1
,
J’.
Proof. Straightforward.
0
As far as simple iterations are concerned, every nested antichain may be regarded a single condition. Namely, nice sequences exist for all nested antichains. This is a main technical lemma stated as follows. 4.10 Lemma. Let J = (Pa I a
5 u ) be any simple iteration with the limit ordinal v. Then for any nested antichian T in J r v , there is x E P, s.t. x is (T,Jrv)-nice.
Proof. x has the only possible definition and we prepare a right induction hypothesis as follows. The harder part is to show x [ a E Pa for limit a 5 v. We must exhibit an appropriate sequence of (J[a)*-dynamical stages for xra. Some details follow. For i < v , we first define .ri E Vpi in four cases as follows. 0 0
+ +
.ri = aor(i l),if i < l ( a 0 ) and aori E Gi,where ao E TO. .ri = b[(i l), if there is (n,a,b) s.t. n < w,a E T,, b E sucF(a), l ( a ) 5 i < l ( b ) and b [ i E Gi.
317
0
~i = l i + ~ if , there is a generic cofinal path (a, I n < w ) through T in V[Gi].(This case explicitly mentioned to indicate the mutual exclusiveness precisely.) ~i = l i + l , otherwise.
The term
lkpi “
ri
is well-defined. We have
~ iE
Pi+l and ~i [i E Gi”.
We next define pi E Pi+l as follows. Here we use the fullness of the iteration at the successors. 0
pi E Pi+l s.t. piri = l i and
Ikpi “ ~ i [ [+i ,1) i
+
pi[[i,i 1)”.
We lastly set z = (pi(i) I i < v). We claim this z works. We show the following by induction on a for a 5 v.
zra
E P,. PI(”): If a0 E TOand Z = min{a,Z(ao)}, then ao[Z x[Z. p z ( a ) : If a E T, and I = min{a,Z(a)}, then arZ 5 zrl. p 3 ( a ) : If a E T,, b E suc$(a), Z(a) 5 a and Z = min{a, Z(b)}, then b[Z E b[l(a)-z[[Z(a),Z). ’p4(a):If t 5 a, w 5~ zrt and w I t p E“there is a generic cofinal path through T in V[G,]”,then w-z[[t, a ) w-1, a).
=
[[r,
Since all but showing ‘po(a)are routine, we provide more on this. We first fix a correspondence ( ( i , u )++ ( & ( i , u )1 k < w ) 1 u E Pa,i< a 5 v for some limit ordinal a ) s.t. 0
(&(i,u) I k < w ) is a sequence of (J[a)*-dynamical stages for u E Pa with Jk(Jr,)*“iyo(i,u)= i”.
This is possible, since P, is, by definition, the dynamical kernel of ( J [ a ) * and we are free t o choose 0-th stages. Let us recall TO= {ao}. We have two cases.
=
5 Z(a0): By induction, we have uo[a zra in (.Ira)*.So the (J[a)*-dynamical stages (&(O,ao[a) I k < w ) work not only for a0 [a but also z [a. We conclude z [a E P,. Case 2. Z(a0) < a: We define (J[a)*-dynamical stages (& I k < w ) as follows. Given any (Jra)*-generic filter G* over V , we form the generic extension V[G*]. Now we first define (the interpretation of) &, by Case 1. a
318
0
cio = l(ao),if a0 E G* [Z(ao).
0
cio = a , otherwise.
We next define (the interpretation of)
&+I
by
- l ( b ) (< a ) , if there is (a,b) s.t. a E Tk, b E suc$(u), Z(b) < a and b[Z(a)E G*[Z(a). &+I = cik+l(#?(a),b[CX)G* ( 5 a ) , if there is ( n , a , b ) s.t. n 5 k, a E T,, b E suc$(a), l ( a ) < a 5 Z(b) and b[Z(a) E G*[Z(a). &+I = a , otherwise.
&+l
0
0
So the basic idea is that we descend through T along the nodes ao, a1 and so forth which are in the corresponding generic filters as much as we can. And once we are about to hit any node whose length is greater than or equal to a , then we cut short the condition at a and switch to the stages which are already associated to the cut one. So if we arrived at some a, and a,+l s.t. a , E T,, a,+l E SUC$(U,), Z(an) < a 5 Z(a,+l), a, E G*[Z(an)and a,+l [Z(an) E G* [Z(an),then we begin to use (&+I @ ( a n )a,+l , [a)I n 5 k < w ) . We claim that these (J[a)*-stages (& I lc < w ) are well-defined and that work for x [ a , so that x [ a E P,. Since the verification of this is quite routine using the induction hypothesis, we leave it to the readers.
4.11 Corollary. Let J = (P, I a 5 v) be any simple iteration with the limit ordinal u. For any fusion structure .F in J, there is a fusion x E P, of
F.
The following generalizes the fullness at the successors. A name of a condition is indeed a condition, so to speak.
4.12 Corollary. Let I = (P, I a < u ) be any simple iteration. If a 5 p < v, p E P, and T is a P,-name s.t. p lkpa (5- E Pp and T[a E G, ”, then there is w E Pp s. t. w [a = p and p [/-pa ‘5-[[a, p) w [[a,p) in P,p ”.
=
+
Proof. We may assume P w < u by lengthening I , if necessary. We construct a nested antichain T in I [ ( P+ w ) so that
. 0
To = { P I . For all b E T I ,b[aIt-p,“b[[a,p) = ~ [ [ a , p ) ”(So . bra decides the value of T , bra 5 Tra and may set b = b [ a - ~ [ [ p).) a,
319
a
T,+1 = T, and suc$(b) = { b } for all n 2 1 and all b E T,.
So we first partitioned p to form TI deciding the values of r. We then just keep attaching the same condition below each element of T, for n 2 1. Let y E be a fusion of T . We may assume y[a = p . Then it is easy to see w = y r p works. 0 5 . Constructions of Fusion Structures
We first quickly review relevant definitions. For a regular cardinal 6 , He denotes the set of all sets which are hereditarily of size less than 8. If a preorder P is in He, then I/-~‘‘{T& 1 T is a P-name s.t. r E H F } =
H r [ G 1 ”An . elementary substructure N of He means ( N ,E) is an elementary substructure of (He,E). We write N 4 He for short. We only consider countable ones in this note. If P E N 4 He, then It-p“{re I r is a Pname s.t. r E N } 4 Hr[G1”. And we write N[G]to denote this c-least
c
elementary substructure M of HrlG1 s.t. N U {G} M . In this section, we prepare for recursive constructions at limit stages of simple iterations. These constructions amount to forming fusion structures. And we have seen that fusions exist for all fusion structures. Here is a typical single step.
5.1 Lemma. (A step forward) Let I = (Pa I (Y < v) be any simple iteration. Suppose we have given (a,i,j,x,(jk(x) 1 k < w ) , D ) s.t. a
a
a E Pl(a),i 5 Z(a) < j < v and j is a limit ordinal. z E Pj has Pj-dynamical stages ( & k ( x )I lc < w ) . u 5 xrl(a) and a-1 IFpj %,(x) = I(u)”. D is a Pi-name and a [ i Ikpi “D is a dense subset of Pij ”.
Then a forces that there is ( y , ( & k ( y ) I k V[Gl(,)]s.t. the following hold. Y E pj, 7J (&k(y)
all lc E u-1 ‘1L
< w ) , u)in the generic extension
5 x, Y V ( a ) E G ( a ) and Y “ i , j )
E
D.
I k < w ) is a sequence of Pj-dynamical stages for y
< w , IFPj %+I(%) 5 & ( y ) ” . S(,)l, ( a. ) I l(u) < j, ‘1L I Y r w IFFj ‘%J(y)
.rl(.>
s.t. for
E G ( a ) and
= Z(u)”.
Furthermore, if 6 is a regular cardinal, N is a countable elementary substructure of He s.t. I € N and a JFq(,)“NU{x, (8k(x) I k < w ) , b , G ~ ( ~ ) }
s
320
" assume (y, (8k(y) 1 k < w ) , u ) E ilk in the above. (We do not require url(a) 5 a , since a would be somewhere out side of M and we want u to be in ilk.)
ilk 4 H,VIG'(a)l hold, then we may
Proof. Straightforward. Use lemmas 2.8 on dense subsets and intermediates, 3.2 on stages and 3.5 on hooking. 0 We present a typical recursive construction at limit stages.
I = (Pa 1 a < u) be any simple iteration. We fix a regular cardinal 6 and a countable elementary substructure N of He with I E N . Assume we have given i and a limit ordinal j with i < j < u. There is no need to assume i , j E N . Suppose we are given ((a,P)H E a p I i 5 a 5 p < j ) s.t. for each a , @with i 5 a 5 p < j , E a p is a Pa-name and the following hold. 5.2 Lemma. (Weaving) Let
1
0
Ikp, "If N
0
w [a E G a }?,. It-p, "If NU {Ga,u}C M 4 Hy[G"l,u E Pp and u[a E Ga,then there is w 5 u s.t. w E E a p ( M )".
U {Ga,/3}C M
< H:[G"], then & p ( M ) E { w
E Pp
(This is going to be realized as an induction hypothesis in the next section).
Now if we start as follows:
a E Pi, x E Pj with a 5 x [ i , (D,I n < w ) is a sequence of Pinames and E HBp' s.t. a 1t-p; YV U {Gi,x} 5 & +I Hy[G'l (and so i , j are in M ) and (D,I n < w ) enumerates the set of all dense subsets of Pij which are members of A?"
M
Then we may construct a* 5 x s.t. a* [i = a and a* forces the following: There is ((an,an,un,wn, M,,x,) 1 n < w) in the generic extension V[Gj] s.t. 0
a. = i, a0 = a , uo = wo = xri, MO= h;rci and xo = x. i < ( Y , < j , a , , u , , w n E P , n , a , ~ ~ n ~ U , anda,EG,,,. N u { G a n } E Mn 3 H;[G"nl and M , E Hy[G"nI
0
x , E Gj n M,, u, decides the value of some 0-th stage for x, to be a,, U , 5 x,[Q, and un E M,.
321
Proof. We combine lemmas 5.1 on a step forward (with M ) and 2.8 on dense subsets and intermediates. We construct a fusion structure T = ( ( a , n ) c) (dazn), (8p.”)I k < w ) ) I a E T,,n < w ) together with (j$f(uin) I a E T,, n < w ) as follows. We first define objects for n = 0 so that 0
I k < w ) and j$f(”olo)
a. = a , d a o l o ) = z, z has Pj-stages
=
M. Since we are free to choose 0-th stages, we may assume
IFPj
= $7
“ @ O N
Since we have I E N 4 He, the above kind of correspondence (i.e., ( i , j , x ) C ) Pj-stages (& I k < w ) with Ikpj “60 = i” for i < j < v, j is a limit ordinal and z E Pj) is available in N . But a Ikpi “NU {Gi,z} C A2 4 ~ ~ [ G, so i we l ~may ~ assume 0
Ikp, LL(6k’o) I k <w) E
a0
We proceed recursively.
( d a t n ) , (8Ftn) 1 k < ~ ) ,
j$f7,.
Suppose we have constructed ( ( a , n ) ++ d (1 ~ a E*T,) ~ )s.t.) for each a E T,, the fol-
lowing hold. 0
0 0
a E Pqa) with i 5 Z(a) < j . E Pj has Pj-stages (8p’”’ I k < w ) . (fusion structure) [Z(a) and a-1 Ikpj “6~’”’ = Z(a)”. (fusion structure) a< -
~
(
~
3
~
)
a I F F l @ )“A2 u {Gq+
Z(a+),
(q“) I k
<
Matn) 4
w)}
H,VIG,,a,l,,
w e construct ( ( b ,n + 1) C ) ( z ( b I n + l ) , ( 6 ~ ~ n +1 1 ) < w ) , A2(bi”+l) ) l b E sucF(a)) t o form T,+1 s.t. for each b E sucF(a), there are u and w with the following. 6, U , w E Pl(b),Z ( U ) 5 Z(b) < j and b
5 w 5 u5
~
(
~
1
”
~
’
rZ(b). )
322
0 0
0 0
z(b’n+l)E Pj has Pj-stages (hr”+” I k < w ) . dbyn+’) < z(a>n). (fusion structure) (fusion structure) Ikpj 6‘ 6k+1 .(a’) < - k bri Ikpie e ~ ( b ~ n[[i, + l )j ) E &”. u decides the value of df’nf’) to l(b). (u-1 lkpj r‘d-f’n+l) brl(a) IkpIca, “z(~J’+’), u E A&’,”’’. ccw E G ( ~ ) ~ ( j$f(w4),,. ~)( b r w
blt-Pl(b) < i & f ( b , n + l ) = j$f(a,n)G
( b )
r[l(.)>
= Z(b)”)
V[G(b)l,,
l(b))l 4 H0
The construction is straightforward via lemma 5.1 on a step forward. We now take a fusion a* of F via lemma 4.10 on the existences of nice sequences for the nested antichains. We may assume a* [i = a . So we are done. 6. The Simple Iterations of Semiproper Preorders
We recall relevant definitions. A preorder P is called semiproper iff for all sufficiently large regular cardinals t9 and all countable elementary substructures N of Ha with P E N , if p E P n N , then there is q p s.t. q lkp“N[G]n w r = N n w y ” . In this case, q is called (P,N)-semi-generic. It is known that club many N ’ s at a sufficiently large regular cardinal 8 suffices in the above. (pp. 102, 483, 484 in [3])
<
6.1 Definition. A simple iteration I = (Pa I a < v) is called a simple iteration of semiproper preorders, if for any a with a 1 < v, 1t-p- “Para+l is semiproper”.
+
Here is our preservation theorem for the semiproper preorders. 6.2 Theorem. Let I = (Pa I a < v) be any simple iteration of semiproper preorders. Then f o r any i 5 j < v, we have [bpi“Pij is semiproper”. More technically, we have the following. Let B be a suficiently large regular cardinal and N be a countable elementary substructure of H0 with I E N . Then f o r any j < v, the following holds. 0
<
For any ( i , z , a , & f ) s.t. i 5 j , z E Pj, a zri and aIkpj “NU {Gi,z} E j$f 4 HBVIGil”, there is a* E Pj s.t. a* 5 5, a * [ i = a and a Ikpi “a*r [ i , j ) is (Pij,&f)-semi-generic”.
323
Since we have the fullness of names for conditions by corollary 4.12, the above is equivalent t o the following.
For any i with i 5 j < v, 1t-p; “if N U {Gi} & M 4 H ,V[GiI, x E Pj n M and xri E Gil then there is y 5 x s.t. y r i E Gi and y[[i,j) is (Pij, M)-semi-generic”.
Proof. We proceed by induction on j < u. If i = j, then there is nothing to prove. We may assume i < j. Now we have two cases . Case 1. j is a limit ordinal:
Suppose we have given (i, x,a , M) as in the statement. We make use of lemma 5.2 on weaving. We first define (I& 1 i 5 a 5 ,Ll < j). For each a and ,L? with i 5 a 5 p < j , we set
II-P~“&/J(M)= {w E PO I wr.: E G a , wr[a,p) is ( P , ~ , M ) semi-generic} for all M with N u {Ga,/3} & M 4 HYIGal (and so Pap E M ) ” . By induction, for any a and /? with i 0
5 a 5 /3 < j, we have the following.
lkpa“If N U {Ga,u}E M 4 Hr‘Gal, u E Pp and ura E G,, then there is w 5 u s.t. w E E u p ( M ) . (i.e., wra E Ga and wr[a,P)is (Pap,M)-semi-generic)” .
Since we assume 0
a
5 xri and a Itp; “ N u { G i , x } & h!f 4 HY[Gi1”.
We may take a sequence (D,1 n 0
< w ) of Pi-names s.t.
a lkpi “(b, 1 n < w ) enumerates the set of all dense subsets of Pij which are members of A?”.
Now by lemma 5.2, we get a* 5 x s.t. a*[i = a and a* forces, among others, that there is a sequence of objects ( ( a n M,, l w,,x,) 1 n < w ) in the generic extension ~ [ G j s.t. l 0
i = a0 5 a, 5 a,+1 < j . 2 G i = M,, M , 4 H~V[GL-,I and M , E
0
Mn C Mn+l = Mn[Ga,an+l]. x = xo 2 x, 2 x,+1 in Pj.
c
H,V[~~~I.
324
Since the D,’s have been taken care of, we may conclude
But we have
Case 2. j is a successor ordinal:
<+
Let j = 1. Notice that since we assume i by induction, we have a* E Pc s.t. 0
< j, we have i 5 <. Now
a*[i = a, a* 5 x[< and a IFpi “a*[[i,<) is (Pie, $if)-semi-generic”. 1Fpc “Gt E
$if[Gic]4 Hr‘G‘l and so Pcc+1,x[[<,J+l) E & f [ G i ~ ] , where $if[Git]abbreviates (G)(bcri)[Gt [[i, <)I”. So we have But a*
a* 1Fpc “There is 7r semi-generic”.
5 x[[<, < + l )in Ptc+1 s.t.
7r
is (Pee+,,$if[Gie])-
Now by the fullness at the successors, we may take a* E P C +s.t. ~ a* [<= a* and a* lFpc “a* [[<, 1) G A”. It is easy to check this a* works. 0
<+
The following would be a working knowledge on the simple iterations of semiproper preorders.
6.3 Lemma. (Weaving after preservation theorem) Let I = (PaI Q < v) be a n y simple iteration of semiproper preorders. W e fix a suficiently large regular cardinal 9 and a countable elementary substructure N s.t. I E N as usual. Let j < Y be any limit ordinal in N . Suppose we have ( F a p 1 Q 5 p < j ) s.t. f o r a n y Q and p with a 5 p < j , the following hold.
325
0
Jtp_ “If N U { G a , D } C M 4 H r [ G a l ,then p a p ( M ) C {w E Pp I w [ a E Ga}”. lkpa ‘YfNU{G:,,u} M 4 H r [ G a l , N n W 1 = M n w l , u E and u [ a E G a , then there is w 5 u s.t. w E M n Fap(M) and so there is v 5 w s. t. v [ a E Ga and v [ [ a D) , is (Pap,M)-semi-generic”.
Then for any x E Pj n N , there is y 5 x s.t. y forces the following: There is ((a,,M,)
I n < w ) in the generic extension V[G:j]s.t.
o=arJ
N = M o , M oU {GaY,,an+l}C Mn 3 Ho“[Ga,I ,M , E H,V[~Q*Iand Mn+1 = Mn[GanLYn+ll, Mn fFl anan+l ( M n )n Ga,+, # 0. ivnnwl =Nnw,. M o [ G j ] n H f C U { M , n H f I n < w } and so Mo[Gj]nwl= Monwl.
. 0
Proof. Similar to lemma 5.2 on weaving with i = 0. We also incorporate elements from theorem 6.2 on the preservation for semiproper. Details are left. We finish this note with the following observation on the chain conditions of simple iterations. We are sort of free to get the chain conditions. 6.4 Theorem. Let n be a Mahlo cardinal and J = (Pa 1 a 5 K ) be any simple iteration. If for all a < n, I Pa I< n, then Pn has the n-C.C.
Proof. We first make two claims and observe that these suffice for showing the n-C.C.We then show these two claims.
Claim 1. Let i L K. be a limit ordinal. For any x E Pi, there is (T,y) s.t. T is a nested antichain in Jri, y E Pi is ( T ,J[i)-nice and y 5 2. Claim 2. Let i 5 n be a regular uncountable cardinal s.t. for all (Y < i, Fix any ( T ,y) s.t. T is a nested antichain in J [ i and y E Pi is (T,J[i)-nice. Then there is a < i s.t. y = y[a-li[[a,i>in Pi. To observe these suffice, let (pi 1 i E A ) be any indexed family of conditions of Pn,where A = {i < n I i is a regular uncountable cardinal and V a < i I Pa I< i}. For some f ( i )< i < g ( i ) < n, we may assume
1 Pa (
Pi = ~ ~ [ ~ ( i ) - i ~ [ [ f ( ~ ) , i ) - P i r [ i , g ( i ) ) - i ~ [ .[I.g ( ~ ) ,
326
By Fodor’s lemma, we may also assume f(i) = f(j)and pirf(i) = p j rf(j) for all i , j E A. So we may find i < j in A with g ( i ) < j . Let p = p i [ j ^ p j [ [ j , n ) . Then p E P, and p 5 p i , p j . Hence P, has the K-C.C.
(8-k
Proof of Claim 1. Since x E Pi, we may fix its Pi-dynamical stages I 5 < w ) . We construct a nested antichain T in J r i as follows. a. E
Fj(a,,),Z(ao)< i , a o 5 xrl(ao) and
This is possible, since x lkp, “8-0
< 3’. Now
acl lkp,“$ = l ( a o ) ” .
suppose we have constructed
T, so that 0
a E Pl(a),Z(u) < i , a
5 xrZ(a) and a-1 IFpi“S,
= Z(a)”.
We want suc?(a) for all a E T,. But a-x[[Z(a), i) Ikpi “Z(U) = d‘, 5 Jn+l” and &,+I is a Pi-dynamical stage. So it is routine t o construct suc$(a). This completes the construction of T . Now let y E Pi be (T,Jri)-nice. Then y Ikp, “there is a generic cofinal path through T in V[Gi]”. Hence ylkp;“if 8- = sup{& I n < w } , then 0 x [8- E Gi and so x E Gi” . We conclude y 5 x.
r8-
Proof of Claim 2. Since i is a regular uncountable cardinal and for all a < i, I Pa I< i assumed, there is a < i s.t. for all n < w , T, C U{fl 11 5 a}. Hence for any ( T ,J ri)-nice sequence y E Pi, we have y y [a^ 1. (We may not have y = yra-1. But it does not matter.) The following is related to remark 3.26 on p. 68 of [5]. 6.5 Theorem. Let n > w1 be a regular cardinal and J = (P, I a 5 n) be a simple iteration of semiproper preorders. If f o r all a < n, 1 P, I< K , t h e n
P, has the n-C.C. Proof. Similar to theorem 6.4. We may prepare the following claim to replace claim 1 in the previous proof. And we may use this claim together with claim 2 as before. Here we set A = {i < n I cf(i) = w1) which is stationary in n.
Claim. Let i < K be a limit ordinal with cf(i) = w1. For any 2 E Pi, there is a < i and a E P, s.t. a 5 x r a and a^x[[a,i) G a n l i r [ a , i ) 5 x. Proof of Claim. We first take a sufficiently large regular cardinal 0 and a countable elementary substructure N of He with x ,P i E N .
327
Since Pi is semiproper, we may then take y E Pi s.t. y 5 x and y is (Pi,N)-semi-generic. Since cf(i) = w1, we may assume this is true n w1 = N n w ~ but ” in N as well. So we have not only y Jkpi“N[G:i] also ylt - pi “ sup( ~[ G :i] n i) sup(^ n i)”. Since x E N , we may also assume that x has Pi-dynamical stages (in 1 n < w) in N . Since 3 5 z and z It-p, “& < i”, we have y lFpi “in E N [ G J n i”. Let us set a = s up( N n i ) < i. Then y I ~ P ~ “ s u ~ { I ~ n~ < w} 5 a”. Since yItpi“xra E G:i[a”,we certainly have yIkpi“x[[a,i) li[[a,i) in Psi". Hence we must have a E Pa s.t. a 5 z [ a and anx[[a,i) a-li[[a,i). 0
=
Note. I would like to thank David Aspero for many comments on this
note.
References 1. H. Donder and U. F‘uchs, To appear in Handbook of Set Theory. 2. C . Schlindwein, Simplified RCS iterations, Archive for Mathematical Logic 32 (1993), 341-349. 3. S . Shelah, Proper and Improper Forcing, Perspectives in Mathematical Logic, Springer-Verlag, 1998. 4. T. Miyamoto, O n iterating semiproper preorders, Journal of Symbolic Logic 67 (2002), 1431-1468. 5. H. Woodin, The Axiom of Determinacy, Forcing Axioms, and the Nonstationary Ideal, de Gruyter Series in Logic and Its Applications 1, de Gruyter, 1999.
328
AN APPLICATION OF N D J ~ R o TO ~ THE CATCH AND THROW MECHANISM
MASAHIRO NAKATA Kobe University 1-1 Rokkodai, Nada, Kobe, 657-8501 Japan E-mail: [email protected]
NORIMICHI SANETO Setoda Highschool Setodachou, Toyotagun, Hiroshima 722-24417 Japan E-mail: [email protected]
MARIKO YASUGI Kyoto Sangyo University Motoyama, Kamigamo, Kita, Kyoto 603-8555 Japan E-mail: [email protected]
Abstract Catch and throw is a mechanism in the programming language dealing with exceptional processes. Nakano in [4,5,6,7] introduced a typing system, named L z v ,that can express the exceptional processes. He proved its soundness by interpreting it by a term (program) system with catch and throw operators. Here we will make use of a system of propositional calculus, named NDJ,,,,, and modify it to obtain a deductive system NDJ,/t as an alternative to L z " . We show that there are maps in both directions between Nakano's term system and NDJ+ that (in a certain sense) preserve reductions. If we regard a term as a program, then this assertion means that a proof and a program directly correspond to each other, and the correspondence is attained by the above maps.
1. Introduction In [4,5,6,7], Nakano defined a term system (named 7 here) which contains the constructs of catch and throw, and he also introduced a typing system,
329
named LZVlthat can represent the catch and throw mechanism. The terms of 7 are supposed to represent programs. Using these, he first gave an operational semantics of I and then proved the soundness of L Z ” by a realization interpretation in terms of 7 . He also reformulated L Z ” in sequent style and proving a cut-elimination theorem. In order to explain the purpose of this article, let us first give an account of the catch and throw mechanism, as well as the use of a connective called exception, at some length. The catch and throw mechanism is a means of non-local exit used in programming languages such as Lisp, C++ and Java. The throw construct throws exceptions and the thrown exceptions are caught by the catch construct. The notion of exceptions in programming languages represents “unusual situations” in computation. Typical exceptions are errors such as the one that happens when one tries to compute When an exception such as the error of division by zero happens, the exception is thrown to the top of the computation tree. If the thrown exception is caught by a catch construct, the catch construct can handle the exception just like usual data. Thus the mechanism of exceptions allows a program to recover from its own errors by itself and is very important in practice. The catch and throw mechanism could be used to handle unusually good situations as well. For example, in computing I17=lf(i), we may stop the computation as soon as we encounter an i such that f (i) = 0, since then the result of the computation must be zero. The logical formula AaB represents the situation “Usually A holds, but, if an exception happens, then B holds.” For example, if f ( i ) = 0 or f ( i ) = i, then the possibility of an “exceptionally good situation” mentioned above is formulated as follows: rI?.lf(i) = n! a ( 3 i . f ( i )= 0 A rIy=lf(z) = 0). The contribution of the present article is to formulate a natural deduction style formal system NDJ+ that corresponds to the term system I, and to define maps between the terms of I and the proofs in NDJ+ that (in a certain sense) preserve reductions. (The result is summarized as Theorem 3.1.) The system NDJ+ is a modification of NDJ,,,,, the constructive propositional calculus formulated in the natural deduction style. The map from NDJ+ to 7 mentioned above is also expected to be useful for proof animation, proposed by Hayashi, in [2] for example. Namely, the validity of a formal (supposed) proof is to be tested by computing the term (program) extracted from the (supposed) proof for some examples of inputs. The map from 7 to NDJ,,, will be also useful in giving a proper proof once
h.
330
a program is corrected (if erroneous). In [13,14,15], Yasugi, Ryu and Nakata proposed and investigated a system of classical predicate calculus in the natural deduction style, named NDK, whose characteristic is that introductions and eliminations of any logical connective can be executed at any place in the disjunctive components of a formula. In this way, the law of the excluded middle can be deduced as a consequence of an inference, not as an axiom. For example, the inference rule A-elimination applies to C V ( A A B ) V D to deduce the consequence C v A v D; the inference rule +-introduction applies t o r V B V A with a live assumption [A] to deduce the consequence r V (A -+ B ) V A with [A] discharged, where I? and A represent zero or more formulae. In [8] Nakata considered the propositional part of NDK, named NDK,,,,, and proposed a variation of the implicational part of NDK,,,,, named NDK’. In [ll],Saneto made some observations about the constructive part of NDK,,,,, named NDJ,,,,, that is obtained from NDK,,,, by restricting the introduction of + or -,t o only the whole formula in the premise. For example, the +-introduction as above is allowed only t o conclude A + (I’v B v A). He then applied it to define a system NDJ,p, which expresses the catch and throw mechanism, modelled after Nakano’s method in [4,5,6,7]. With NDJc/t, the correspondences between a term system and a logical system immediately follow. The map from the deductive system NDJClt to the term system ‘T is modelled following Nakano’s idea, while the converse map is modelled after [12]. This article has emerged from Sections 6 and 7 of [ll]and also from [9]. In Section 2 of this article, we introduce the system NDJ+; in Section 3, we first review the definition of the term system 7, and then define correspondences between terms of ‘T and proofs in NDJ+ (Propositions 3.1 and 3.2, Theorem 3.1). At the end, in Section 4, Nakano’s example of realizing the catch and throw mechanism is presented in our language. Related work can be seen in [3]. A comprehensive report on the propositional subsystems of NDK has been presented in [9]. We have followed the notations of [l]and [lo] for systems of natural deductions.
2. A system NDJ+ representing catch and throw We now define the system NDJ+. NDJ@ a modification of NDJ,,,,, a formulation of the constructive propositional logic, t o a system that can express the catch and throw mechanism. In the conclusion of an inference of NDJc/t, there occurs a formula of the form A V AT. Here T represents a
331
set of assignments of tag variables. Tag variables, a new logical connective a and new rules of inferences are added to NDJ,,,, based on Nakano's ideas ([4,5,6,7]).
Note Although the logical connective 1is employed in NDJ,,,, (being the usual propositional calculus), we do not need it for our purpose as it does not represent any computational operation and hence we do not include it in the system NDJ+. Definition 2.1. 1) The propositional formulae are defined as usual using the connectives A,Vand -+ as well as round brackets, although these will be omitted in order to avoid notational complication, unless confusion is likely. 2) A new logical connective a will be introduced, so that, if A and B are formulae, then A a B is also a formula. It is called exception by Nakano. (a is a kind of disjunction, details of which have been explained in the introduction.) 3) Tag variables are denoted by u, v,. . .. A formula B with an assignment of a tag variable u is written as B". B and B" are identical as propositions, but their roles are different. A formula with a tag assigned is called an error message. 4) The disjunctive components of a formula A are defined as follows. 1. If A is an atomic formula, then the disjunctive component of A is A itself. 2. If A is of the form BVC,then disjunctive components of A are the disjunctive components of B , the disjunctive components of C and A itself. 3. If A is of the form BAC,B ---f C or BaC,then the disjunctive component of A is A itself. For example, if r = (AA B)V (CV D)where A , B, C and D are atomic, then the disjunctive components of are I?, A A B ,CVD,C and D. 5 ) Formal proofs in NDJc/t will be defined below. Labels of assumptions will be denoted by z, y , . . .. Signs such as (z),( x ,y) written at the left of the name of an inference rule will indicate the label(s) of the discharged assumption(s).
We note that, in an assumption of an NDJ,+-proof as defined below, no error message occurs. In the conclusion of an inference, a formula of the
332
form A V (BY' V BY2 V . . . V BEn) may occur. Here, n 2 0, and we may abbreviate a disjunctive formula with tags such as (BY' V BZ2 v . . . v BXn) by A T . In A V AT, A will be called the main part, and AT will be called the exceptional part. Prior to the definition of proofs, we define the set of substitution contexts, written sCont. Let tFormula denote a tagged formula. sCont*, the extended substitution context, is first defined, introducing an atomic symbol a. sCont* ::= a 1 tFormula I sCont* A sCont* I sCont* V sCont* I sCont* ---f sCont* 1 sCont* a sCont+.
A substitution context, sCont, is defined to be an extended substitution context in which the atom Q does indeed occur. For an sCont, say E T , we write ET[E"] for the result of substituting a tagged formula E" for all occurrences of the atom a in ET. It is a tagged formula. E T [ ] will denote the tagged formula obtained from ET by deleting Q together with the connectives that introduced Q in the construction of ET. The formula E T [ ] is a tagged formula. For example, if ET is X" V (Q vY"), then ET[E"] is X u V (E" V Y") and E T [ ] is X" V Y". Proofs in NDJ,/,: A proof in NDJ+ is built from assumptions and inference rules as listed below. [Assumption] [ x : A] will denote an assumption A (a formula without tag variables) with label x. [Inference rules of NDJ+] In the following, AT will denote a tagged formula, while ET will denote a substitution context.
[x: A] AVET[] (weak) A v ET[Eu]
EvST[] A v ET[E"]
(throw)
sT
A V [A"] (catch) AvET[]
333
(A1 A A 2 ) V AT (AiE) Ai V AT
AV'AT BV'AT ( A A B ) v aT
V
A~ AT (A1 V A2) V AT
(VJ
(Z
= 1,2)
(i = 1,2))
[x:A]
[y : B]
[x: A]
A v ET[E"] ( A a E ) v ET[ ]
(AaEjvET[] A v sT[E"] (aE)
(4
A v AT AT (Contraction) A V A ~ A in the conclusion of throw is an arbitrary formula; AT in the conclusion of (-+ I) is arbitrary, while the exceptional part in the premise of (-+ I) is empty. Definition 2.2. (Reduction.) We define reductions of proofs in NDJ,lt. The reductions we employ are the so-called innermost reductions. In the following, V and M will denote parts of proofs to which reductions no longer apply; M' will denote a simulation of M , that is to say that M' is obtained from M by replacing each formula D in M by D V AT for an appropriate taggcd formula AT. Similarly for Ml. then M' is For example, if M is AA
B
,
( A A B ) v nT BvAT
334
Now here are the reductions.
:v AvET[] (throw) A v ET [A"] (catch) AvZT[]
:v ==+
AvZT[]
:v EvET[] (throw) A V ET[E"]
:v
:M E V ET [E"] (catch) EVET[]
*
EvET[] (throw) E V ST [E"] (catch) EVET[]
:v
:v
[x: A] : M
A V. AT : M' 3
c vCATV AvTaT (Contraction)
[x: A] : M
:v
EvET[] (throw) A v ET [E"]
:v EVETI1 ===+
I . .
335
Notice that a common tag variable u occurs in the premise of in the conclusion of (aE).
(GI)
and
:v EvzT[] (throw) A V ET [E"] A a E v E T [] ( 4 4 (aE) A v zT[E"]
:v EvET[] (throw) A v ET [E"]
Notice the change of tag variables.
:v Ai
AT : M,!
(i = 1 , 2 )
:v
Let II denote a proof in NDJ,/t, and let II[IT'] denote the fact that II' is a subproof of II,that is, IT' c II and II' is itself a proof. II[II,] is said to be reduced t o II[II2] (written II[II,] =+ I I [ I I 2 ] ) if IT1 ==+ I I 2 . In general, for proofs II and II*,we write II =j 11*when II is reduced to II* in the sense above.
336
3. NDJ,/t and the term system We first introduce Nakano’s terms and their reductions (cf. [4,5,6,7]), and then give a term-representation of NDJ+ using these terms. A converse map will then be defined. 3.1. Review: Nakano’s terms and reductions
Definition 3.1. Three mutually exclusive sets of symbols are assumed. Const : a set of individual objects c, d , . . . Var ! : a countable set of individual variabIes 2,y, z , . . . Tvar ! : a countable set of tag variables u, v, w ,. . . Definition 3.2. Terms of the term system T are defined as follows. Term ::= Const I Var I let Var = T e r m . Term I throw Tvar Term I catch Tvar Term 1 X Var . Term* 1 Term Term I ( Term , Term ) I proj1 Term I pro5 Term I injl Term I inj2 Term I case Term Var . Term Var . Term I K Tvar . Term I Term Tvar Here Term* denotes a Term that has no free occurrences of tag variables. M , N , K , L, . . . will be used to denote terms. A Term that has no free occurrences of variables will be called a closed term. Definition 3.3. The set of values, Val, will now be defined. Val ::= Const I Var 1 X Var . Term 1 ( V a l , Val ) I injl Val I inj2 Val I K Tvar . Val I K Tvar . throw Tvar Val V ,V‘, W,W’, . . . will be used for elements of Val. Definition 3.4. The set of term-like expressions with an atom *, Cont, will be defined next. A member of Cont is called an evaluation context, and * is called a hole. Cont ::= * I let Var = Cont . Term I throw Tvar Cont I catch Tvar Cont 1 Cont Term I Val Cont I ( Cont , Term ) I ( Val , Cont ) I proj, Cont I proj, Cont
337
I injl Cont I injz Cont I case Cont V a r . T e r m V a r . T e r m I K Tvar . Cont I Cont Tvar The result of filling all the occurrences of the hole a context C will be denoted by C [ M ] .
* with a term M
in
Definition 3.5. Call-by value rewriting His defined by the rules below.
C denotes any evaluation context which is not * itself. V denotes a value . M [ V / x ]denotes the result of substituting V for x in M . catch u V H V catch u C[throw u V ] catch u (throw u V ) catch u [throw u V ] H V let x = V . M M[V/x] (A x . M ) v M[V/x] K u . C[throw u V ] H K u . throw u V
---
( K U . V ) V
v
u . throw u V ) v H throw v V v Projl (V , W ) W PrOjz (V , W ) case ( i n j l V ) z . M y . N M[V/x] case (injz V ) x . M y . N H N [ V / y ] C [ M ]is said t o be reduced to C [ N ]if M N , and this fact is written (K
-
as C [ M ]
C"].
A call-by value rewriting expresses the innermost computation. 3.2. Assignment of terms t o NDJ,.t-proofs We will assign a term from I to each proof in NDJc/t. Given a proof in NDJ,/,, say II, then
:11 t:CkAT will express the fact that the term t is assigned to II. t is meant to represent the computation process of deducing the main part, C , of the conclusion of the proof II. Proposition 3.1. Let II be a n NDJ,/,-proof (see Definition 2.1), and let r(II) denote the assignment of a term from I to II as defined below in Definition 9.6. Then r preserves the reduction. That is to say, suppose
338
-
~ ( n i'f)~ ( I 2i)s not a value; ~ ( n=) ~ ( 1 1 ' if) ~ ( nis )a value; -r(II)i s a value i f no reduction applies to II. II 3 II'. Then ~(n)
Definition 3.6. (Assignment of terms by x is assigned t o
T.)
[x : A] (assumption) That is, x : [ x : A ] ,or T ( [ X : A ] )= x. The label x is regarded as a term (variable) deducing the assumption A. For a compound proof II,~ ( n is )defined inductively according to the rule of inference 5) in Definition 2.1.
[x : A]
N : A V AT M : C VAT (let) let x = N . M : C v AT
t : A V ET[ ] (weak) t : A V ZT[E"] M : E V ET[ ] (throw) throw u M : A V ZT[E"]
M : A v ZT[A"] (catch) catch u M : A v Z T [ ]
[x: A]
xx
M': B . A~ + B
v nT
M : A ~ B V A TN : A V A T (+El
I)
(4
M N : B V A ~
A'
M : A B v AT (A l E ) projl M : A v AT
[ x : A]
M : A A'B v AT proj, M : B V AT
(A2E)
[Y : BI
L ~ A V B V A TM : C V A T N : C V A T( v E ) case L x.M y.N : C v AT
339
It is routine work to prove the proposition. 3.3. Assignment of proofs t o terms Conversely, one can assign a proof to a term. Suppose first, that a map ,B from variables to formulae is arbitrarily chosen with p ( x ) = A .
Definition 3.7. d ( x ) = [z : A] if p ( x ) = A.
d(1et x = N . M )
=
[x : A] d(N) d(M) AvAT C v A T (x) CvAT
d(M) d(throw u M ) = E v Z T [ ]
A v ZT [E"]
d(X2.M) =
d ( i n j i M )=
[x: A] d(M) lIvAT
( A+ B ) v aT
d(catch u M ) =
(x) d ( M N )
Ai v AT (A1 V A2) v AT
d(M) A v sT[A"] AVZT[]
d(M) d(N) (A+B)VAT AvAT
-S
B V A ~
340 [ z : A]
d(L)
d(M)
(AvB)vAT CvAT CvAT (x,Y 1 CvAT
d(case L x.M y.N) = d(M)
d(Ku.M) =
[ y :B ] d(W
A v ET [E"] ( A a E )V E T [ ]
d(W
d(Mu)=
( A a E )v s T [] A V ET [E"]
Proposition 3.2. Given an arbitrary map p from variables to formulae with P(z) . , = A. Then the assignment d , given in Definition 3.7, of a n NDJ,+-proof to a term relative to /3 such that d preserves reductions, that is, i f t Hs, then d(s). d(t)
*
The proposition can be proved by examining the definitions. We now sum up our results.
Theorem 3.1. The NDJ,/,-proofs and I - t e r m s are related as follows. such that the comThere is a map from NDJ,/,-proofs to 7 - t e r m s , T(II), putation of 4 3 ) is reflected in the reduction of TI, and, i f ~ ( r Iis) a value, then the reduction of ll preserves the same value. Conversely, there is a map from 'T to NDJ,/,, d ( t ) , such that d preserves reductions. 4. Realization of catch and throw: an example
We will explain how to realize the catch and throw mechanism in terms of the term assignments to NDJ,/t-proofs using an example from 1.2 of [7]. Let P be a program which, for a sequence of input data satisfying specifications A l , . ,. ,A,, either outputs data satisfying a specification C or returns an error message E . Let us write this condition of P by:
P : A 1 ,..., A , H C ; E Let P* denote a program that does the following task. Given programs Q, R, S such that
Q : A 1 ,..., A , H D ; E R : A1,. . . ,A,, D
t--)
F;
S : A1 . . . ,An, E
t-+
F;
341
P* first runs Q. If the output satisfies D ,then it proceeds to R, and if the output satisfies E , then it proceeds to S. P* can be represented in NDJ,+ as follows, where [A] abbreviates [ A l l . .. [A,]. (We will omit the parentheses enclosing a term and the corresponding main part of a formula.)
throw u ( l e t x = Q.R) : E v E" v F" SIF catch u (throw u (let x = Q.R)): E v F" throw v S : F V F" (Y> let y = (catch u (throzo u ( l e t x = Q.R))).(throwu S ) : F V F" catch u ( l e t y = (catch u ( t h r o w u ( l e t x = Q . R ) ) ) . ( t h r o w u S ) ) : F
Acknowledgement The authors are grateful to Hiroshi Nakano and Susumu Hayashi for supplying us with knowledge and information about catch and throw mechanism. We are also indebted to the referee, according to whose penetrating comments and advice the manuscript has been revised. References 1. S. Hayashi, Surironrigaku (Muthematical Logic), in Japanese, Korona Pub.Co., 1989. 2. S.Hayashi, R.Sumitomo and K.Shii, Towards animation of proofs-testing proofs b y examples, TCS 272, 177-195, 2002. 3. Y.Kameyama and M.Sato, Strong normalizability of the non-deterministic catch/throw calculi, TCS 272, 223-245, 2002. 4. H.Nakano, A constructive formalization of lhe cutch and throw mechanism, Proc. 7th LICS, 82-89, 1992. 5. H.Nakano, A constructive logic behind the catch and throw mechanism, APAL, 69, 269-301, 1994. 6. H.Nakano, The non-deterministic catch and throw mechanism and its subject reduction property, Logic, Language and Computation, LNCS 792, 61-72, Springer, 1994. 7. H. Nakano, Logical structures of the catch and throw mechanism, PhD Dissertation, University of Tokyo, 1995. 8. M.Nakata, Classical system NDK and its propositional subsystem NDKp', Master's Thesis (in Japanese), Kyoto Sangyo University, 1998.
342 9. M.Nakata, N.Saneto and M.Yasugi, Classical and constructive propositional subsystems of NDK (in Japanese), Bulletin of the ICS, Kyoto Sangyo University, 15, 37-81, 1999. 10. D. Prawitz, Natural Deduction, Almqvist and Wiksell, 1965. and system Nc,t which realizes 11. NSaneto, Constructive system NDJp,,, catch and throw, Master’s Thesis (in Japanese), Kyoto Sangyo University, 1998. 12. M.Yasugi and S.Hayashi, Interpretations of transfinite recursion and parametric abstraction in types, Words, Languages and Combinatorics I1 (edited by M. Ito and H. Jurgensen), 452-464 ,World Scientific Publ. Co., Singapore, 1994. 13. M.Yasugi and M.Nakata, O n natural proofs in NDK (in Japanese), Proceedings of RIMS, 976, 13-26, 1997. 14. M. Yasugi and M.Nakata, NDK and Natural Reasoning, Proceedings of the 6th ALC, 285-309, 1998. 15. M. Yasugi and K. Ryu, NDK, A New Classical System, Bull. of the Inst. of Computer Science of Kyoto Sangyo University, 11, 1-25, 1994.
343
THE CURRY-HOWARD ISOMORPHISM ADAPTED FOR IMPERATIVE PROGRAM SYNTHESIS AND REASONING
IMAN POERNOMO AND JOHN N.CROSSLEY School of Computer Science and Software Engineering, Monash University, Australia E-mail:{ ihp,jnc}@csse.monash.edu.au
Abstract The Curry-Howard isomorphism permits the representation of intuitionistic logic as a constructive type theory. It has often been exploited in the implementation of interactive theorem provers. It also forms the basis of the proofs-as-programs paradigm, an approach to the synthesis of functional programs from intuitionistic proofs (see e.g., [l,2,3, lo]). In this paper, we present a constructive version of a Hoare logic (cf. [9]) for reasoning about imperative programs t o which the Curry-Howard isomorphism may be adapted. This allows us to take advantage of the isomorphism for theorem proving implementation, and for the synthesis of correct imperative programs with return values, following the proofs-as-programs paradigm. 1. Introduction
It is well-known that intuitionistic logic can be presented as a constructive type theory where proofs correspond to terms, formulae t o types, logical rules correspond to type inference and proof simplification corresponds to term evaluation. This property, known as the Curry-Howard isomorphism, is often used in the implementation of interactive theorem provers. Amongst other advantages, the isomorphism permits the automation of proof simplification, and complex proof tactics and parametrized lemmata to be given simply as functions over terms. Various authors have used the Curry-Howard isomorphism for the synthesis of functional programs from constructive proofs [1,2,3] - an approach known as the proofs-as-programs paradigm. The idea is as follows. A description of the required program behaviour is given as a formula,
344
P , for which an intuitionistic proof is given. The constructive nature of the proof allows us to transform this proof into a functional program whose behaviour meets the requirements of P. For example, the formula Vx : Int.Vy : I n t . 3 ~: Int. gcd(x, y , z ) can be interpreted as a specification of a program that takes x and y as input values, and returns as output value z , the greatest common divisor of x and y . An intuitionistic proof of this formula enables the construction of a witness term for z . The proof is then transformed into a functional program f satisfying the formula as its specification. In most modern approaches to proofs-as-programs, the proof is represented as a term of a constructive type theory, with the required program given as a term of a simply typed lambda calculus (such as a subset of the programming language ML). The transformation is given as an extraction mapping from the former type theory to the latter. This paper provides a constructive version of a Hoare logic (see [9]) for reasoning about imperative programs, to which the Curry-Howard isomorphism and a proofs-as-programs approach may be adapted. Theorems of the logic consist of an imperative ML program and a formula that describes the program’s side-effects (see below, Section 2) by means of pre- and postconditions. Our logic uses a natural deduction proof system. Therefore our calculus can be understood as a constructive type theory where proofs correspond to terms and formulae to types. We can give an extraction map from proof-terms of this type theory to programs (in SML). This enables SML programs to be synthesized from proofs of their specifications. This approach to synthesis adapts approaches to intuitionistic proofs-as-programs synthesis. In particular we adapt our work from [14] and [lo] and Berger and Schwichtenberg’s from [l], where a notion of modified realizability defines when a program of a simply typed lambda calculus satisfies a intuitionistic formula as a specification. Note that a theorem in the Hoare logic consists of an imperative program and a truth about the program. Thus, the logic can already be used to synthesize programs. However, such programs do not involve return values. We adapt modified realizability to specify, prove properties about required return values, and an extraction map to synthesize required return values from proofs in the logic. The advantage of our adaption is that the user need not think about the way the return value is to be coded, but can still reason about and later automatically generate required return values. For example, given a
345
constructive proof in Hoare logic of the form program s := 10
0
truth such as
3z.Even(z)
we shall be able to synthesize a return value f in a program of the form s := 10;f
where the function f (with program f ) is a side-effect-free program (such as !s * 2) that realizes the existential statement and acts as a witness for the z. (Here s is the argument o f f and choosing !s * 2 guarantees that we get an even number computed.) The user should have no need to manually code the return value, but instead works within the Hoare logic to prove a theorem, from which the return value is then synthesized. This paper proceeds as follows: In section 2 we present some preliminary syntactic and semantic assumptions about imperative programs. In section 3, we explain how the formulae of our calculus describe side-effect and return values of imperative programs. Our calculus is presented in section 4. Section 5 outlines how we adapt the Curry-Howard isomorphism and to our calculus, and section 6 shows how proofs-as-programs-style program synthesis can be achieved using this adaption. We also outline how these results are implemented and provide a small case study illustrating our work. We briefly review related work and provide concluding remarks in section 8. 2. Preliminary definitions and assumptions
Our results are given with respect to a set StateRef, whose elements denote possible SML state references. We consider a subset of SML programs, SML(StateRe8, whose only state references are elements of StateRef. In this subset, the state references of StateRef can be treated as global, and we assume their declaration as implicit, prior to executing code. For instance, if r € StateRef, then we also have r :=!r + 1;!r; ; € SML(StateRe3, where we assume that r has previously been given a declaration of the form l e t r = ref 0. We distinguish the side-effect-free programs of SML as Pure ML. These programs involve a lambda calculus and may use dereferenced state references (of the form ! r T € StateRef), but cannot change the value of state references. Our results assume a set of pre-defined SML programs, Prog. Our calculus is used t o build larger programs from the basic constructs of SML and this set of (black-box) programs.
346
Our results are parametrized with respect to a many-sorted algebraic signature AVT, consisting of sorts and sorted constant, function and predicate symbols. Terms of AVT (for abstract data type) are formed as usual using constant and function symbols over a distinguished set of variables VaTdvr. We assume that every closed term of AVT, Closed(AD7) can be executed as a side-effect-free (and state-reference-free) program of SML, and each sort is a type of SML. Elements of CZosed(AVT) form the evaluated return values of our programs. We use typewriter font to denote when a AVT term is to be used in an S M L program, and Times Roman font when terms axe used in our calculus. For our present purposes, we take AVT to include a simply typed lambda calculus over the basic data types of S M L with appropriate predicates, and require that we have disjoint union types: type ‘a’b disjointUnion=Inl of ’a I Inr of ‘b;;. Given a first-order many-sorted formula P formed from A D T , and an interpretation L of the variables V a r d v r , we write 11, P when P is true in the intended model for AVT under L. Because dVT terms are S M L programs, the intended model of AVT can be defined using the operational semantics of AVT terms in SML. For instance, for any L, It, Vx : int. ( f u n y => y)(x) = x holds, because, for any S M L integer x, the lambda application (fun y => y)(x) is equal to x. We represent the state of a program’s execution by a collection of value assignments, one for each state reference of StateRef. Formally, we define the set of states C to consist of functions, a : StateRef --+ ADT. Each 0 E C represents a possible state of the computer’s memory, where a ( . ) is the value assigned to state reference s which is in StateRej For example, if is the state after executing r := 40, then a ( . ) = 40,because we know that r stores an integer reference ref 40. A side-effect is the result of executing an imperative program: a transition from an initial state of a computer’s memory to a final state. Because SML(StateRe8 programs only use global state references from StateRef, we can define a simple semantics for representing the side-effect of a program’s execution. Given an SML(StateRe8 program, we write (p,a) 6 (1,a’) to mean that, starting from an initial state a, the program p will terminate and result in the value 1 and a final state 0‘. For the purposes of this paper, we take 6 to be the transitive closure of the relation D given by the operational semantics presented in Fig. 1. For instance, if r E StateRef, then (r :=!r 1;!r;;,0) 6 (1,a’) holds,
+
347
a(b) --+AD7 true
( i f b then p else q, a) D (p, a)
(condl1
false
a(b)
( i f b then p e l s e q, a)D (q,a)
(cond2)
u(b) - ) A D 7 f a l s e (whilel) (while b do c; done, a) D ((>,a)
(c, a) D (r,u‘) (while2) (while b do c; done, a) D (while b do c; done, a’) a(b)
true
Figure 1. Operational Semantics for our programs.
when, for instance, a(.) = 14 and a’(r) = 15. The execution of an imperative program will produce a whole range of side-effects, depending upon the initial state of the memory. We formally denote such a range by a set-theoretic relation, a side-effect relation, between initial and final states. The set of side-effect relations, R e l , is defined as a subset of the power set P(C x C) of pairs from the set of states C. As usual, we write a R a’ if (u,u’) E R. A side-effect relation provides a semantics for an imperative program when the relation’s state pairs consist of the possible initial and final states for the program.
Definition 2.1. Side effect relations. A side-effect relation R is the semantics of an SML(StateRefi program p, written R = [p], when, for all
348
states IT, IT'E 2, (T
* (p,
R IT'
IT)
6 (*, IT') for some value *.
It follows from this definition that if aRa' and ITRO", then is uniquely determined by the initial state c.
IT'= IT",
i.e.
(T'
3. Specification of side-effects and return values
Our logic is used to specify and reason about two aspects of imperative program behaviour: possible side-effects and possible return values. Possible side-effects are specified as pre- and post-conditions (in a single formula style (as opposed to the semantically equivalent Hoare-triple style of [9]). Because a program's side-effect is described in terms of initial and final state reference values, prior to and after execution, these initial and final state reference values are respectively denoted by the name of the state reference with a ()i and with a ()f subscript. For instance, the formula rf > ri describes every possible side-effect of the program r :=!r 1. It specifies side-effects whose final value of r, denoted by r f , is greater than the initial value, denoted by ri. Possible return values are specified as the required constructive content of a formula, in the same way that functional programs are specified in constructive type theories. So, for instance, the formula 3y : int.Prime(y)A y > ri describes a return value y of a program, such that, for any side-effect, y is prime and greater than the initial value of r.
+
3.1. Formulae The formulae of our basic (intuitionist) logic are defined as usual. Our quantifiers are sorted, and take sorts from ADT. For instance, if int E AVT, then 3x : int.x = 0 and Vx : int.x = x are well-formed. To enable the specification of side-effects, our formulae involve terms of ADT, extended over the subscripted StateRef symbols (which denote initial and final state reference values). For instance, if r E StateRef then ri * 20 r f is a wellformed term that may be used in our logic.
+
3.2. Specification of side-eflects
In order to define when a formula is true of a program's execution, we define when a formula is true of a side-effect relation from Rel. Take a formula P of our calculus. Let IT and IT' be initial and final states for some side-effect
349
relation. We write P,"' for the formula formed from P by replacing every initial state reference value symbol si (s E StateRefi by an actual initial state reference value u ( s ) , and similarly for final state references. Then, given a relation R E Rel, an initial state u and a final state u' such that u R u', then we write R It: P when R It, P,"'. Definition 3.1. A formula P is true of a side-effect relation under the interpretation L R, written R It, P , when (a,~') ER
+ R It-, P,"'
A formula is true of a program p under the interpretation L if it is true of the relation for the program for L: that is, when [p] It, P. When this holds for every 1, we write [p] It P. 3.3. Specification of return values
It is also possible t o use formulae to specify possible return values of imperative programs. This is done by extending the way that formulae of intuitionistic logic specify a functional programs, according to the proofsas-program paradigm. To understand this form of specification, we first need some definitions.. We first need to define Harrop formulae. Definition 3.2. Harrop formulae, see [8]. A formula F is a Harrop formula if it is 1) a n atomic formula, 2) of the form ( A A B ) where A and B are Harrop formulae, 3) of the form ( A + B ) where B (but not necessarily A) is a Harrop formula, or 4) of the form (Vx : s.A) where A is a Harrop formula. We write Harrop(F) if F is a Harrop formula, and THarrop(F) if F is not a Harrop formula.
We also need to define a sort-extraction map xsort which extracts the sort from formulae to sorts of AV'T. This is given by the axioms of Fig. 2. We can now define the Skolem form of a program-formula pair. This is the usual definition for first-order many-sorted logic over d V 7 (e.g., that of [l]).However, because our formulae are first-order many-sorted logic over AV'T extended by state identifiers, the definition can be used for our formulae. As usual inl(z) and inr(x) are the first and second encodings of x in pairs and fst(y) and snd(y) are the first and second components of the pair y.
350
F
AAB AVB A+B Vx : S.A 32 : S.A
I
xsort(F)
Unit if not Hurrop( B ) xsort ( A ) xsort(B) if not Hurrop(A) xsort(A)* xsort(B) otherwise xsort(A)Ixsort(B) xsort ( B ) if not Hurrop(B) xsort(A)-+ xsort(B) otherwise s -+ xsort(A) if Harrop(A) s * xsort(A)otherwise Unit
{
{ {"
Figure 2. Inductive definition of xsort(F),over the structure of F, where P is an atomic predicate.
Definition 3.3. Skolem form and Skolem functions. Given a closed formula A, we define the Skolemization of A to be the Harrop formula Sk(A)= Sk'(A,@),where Sk'(A,AV) is defined as follows. A unique function letter f~ (of sort xsort(A))called the SkoZem function, is associated with each such formula A. AV represents a list of application variables for A (that is, the variables which will be arguments of f ~ ) .If AV is (21 : s1,.. . ,x, : s,} then f ( A V )stands for the function application
Wdf, ( x 1 , .. .,4 ) . If A is Harrop, then Sk'(A,AV) If A B V C , then
= A.
Sk'(A,AV) = (Vx : x s ~ r t ( A ) . f ~ ( A =V inZ(x) ) -+ S l c ' ( B , A V ) [ x / f ~ l ) A ( b : xsort(B).f~(AV) = inr(y)-+ Sk'(C,A V ) [ y / f c ] ) If A
B A C , then
351
In the proofs-as-programs approaches of [l]and [14], a formula A specifies a functional program p if, and only if, the program is an intuitionistic modified realizer of A , now defined. Definition 3.4. Intuitionistic modified realizers. Let p be a term of CZosed(dD7) (and so, by the assumptions of Section 2, p.3, p can be considered to be a functional SML program that does not contain any state references). Let A be a first-order many-sorted formula predicating over A D 7 (but not using state identifiers). Then p is an intuitionistic modified realizer of A under the interpretation L when
Recall that, as discussed in Section 2 (p.3), the elements of Closed(dD7) are the return values of SML programs. So we can use the definition of intuitionistic modified realizability to define how our formulae specify return values. Definition 3.5. Return-value modified realizer. We say that an SML program p is a return-value modified realizer of a formula A under the interpretation L , when for for every n, n' such that (P,4
6 (a,4
the SML program a is an intuitionistic modified realizer of A,"' under this case, we write p retrnr, A. If p retrnr, A holds for every L , we write p retrnr A.
L.
In
A formula A of our logic specifies the return value of an imperative program p if p retrnr A. As a simple example, the Skolem form of the formula A z 3y : int.y = r i 1 is f~ = ri 1. It is true that, for any initial state n, (r :=!r + I;!r,0)C; (n(r)+i,n'). Also, ( f = ~ ri 1):' is f~ = o(r) 1.
+
+
+
+
352
I-1.A
(Axiom-I) if A E AX (assign) where s E StateReL
I- s := v 0 sf = tologici(v) I- p
0
c)
(tologic;(b)= true -+ k q 0 (tologici(b) = f a l s e -+ I- if b then p else q C
c) (if-then-else)
t- p (A[si/G] -+ B [ S f / B ] )
9 0 (B[Bi/G] -+ C[Sf/V]) (sed I- p; q 0 A[Si/G] -+ C [ S ~ / B ]
where A and B are free of state identifiers. I- q 0 (tologici(b) = true A A[si/v]) -+ A[sf/Uv] E A[&/v]-+ A[Sf/U] A tologicf(b) = fa lse
(loop)
where A is free of state identifiers.
F p o P I-l,,P-+A tpoA
(cons)
Figure 3. The basic rules of our calculus. Intuitionistic deduction is given in Fig. 4 based on the axioms Az(dDD7).
So, for every initial state (T,o(r)+l is an intuitionistic modified realizer of A. In this way, the formula A can be interpreted as a specification of an imperative program that returns an integer value equal to the value of r (prior to execution) plus 1. 4. The Calculus
Assertions of our calculus are of the form POA
consisting of a program p and a formula A. The formula is to be taken as a statement about the side-effect relation associated with p, provided that p terminates. Our calculus is a version of Hoare logic, providing a natural deduction system for constructing new program/formula pairs from known program/formula pairs.
353
x is not free in A
c : T is a valid sort inference
Figure 4. The basic rules of many-sorted intuitionistic logic, Int, ranging over terms with initial and final state identifiers. We assume that z , y are arbitrary d V 7 variables of some sort T, and that a is a term of arbitrary sort T. The map tologic takes programs to their logical counterparts if they give a Boolean value (see Def. 4.1).
The basic Hoare logic rules are presented in Fig. 3. It can be seen that each of these rules is equivalent to a rule of the more common presentation of the Hoare logic that uses Hoare triples instead of program/formula pairs - see [13] for a proof of this. The Hoare logic rules allow U S to build new programs from old and to prove properties of these new descriptions (cf. [9,4]). For example, the conditional (if-then-else) rule allows us to build a conditional program from given programs, and to derive a truth about the result. Hoare logic is defined with respect to a separate logical system [4]. Usu-
354
ally, this is classical logic (but we use intuitionist logic). The Hoare logic ~ and the logical system interact via a consequence rule. Assume M t - N denotes provability of formula N from formula M in the separate logical system L. Then the consequence rule can be formulated as follows
t- w p p
t -L
A (cons)
t-woA The meaning of this rule is that, if P implies A in the system L , and P is known to hold for the program w, then A must also hold for w. In this way, the separate logical system is utilized to deduce new truths from known ones about a given program. For the purposes of extending known results on program extraction, we define a Hoare logic that takes intuitionistic logic as its separate logical system. The standard rules that define our intuitionistic logic are given in Fig. 4. In these, in order to go from the programs to the logic we require the map tologici, which transforms an SML boolean function b into a boolean term, for use in formulae. The map replaces all state identifiers of the form !s with initial state identifiers of the form si.
Definition 4.1. Given any term b, we define tologici(b) = b[si/!s] where !8 is every dereferenced state reference in b, and 8i the corresponding list of initial state identifiers. We also define tologicf(b) = b[Sf/!S] where 5 is every dereferenced state reference in b, and Sf the corresponding list of final state identifiers. For the purposes of reasoning about an intended model, our calculus can be extended with axioms and schemata (including induction schemata). For the purposes of this paper, we do not deal with schemata. We shall assume a set of axioms A X , consisting of 1) program/formula pairs satisfying p 0 A E AX if, and only if, [p] It A in the intended model and 2) formulae, so that A E A X if, and only if, It A.
Remark 4.1. In [13] we give a proof of soundness for our calculus over SML execution traces. See [4] for proofs of soundness and a form of relative completeness for a wider range of models.
355
5. Adapting the Curry-Howard isomorphism Our calculus forms a type theory, LTT, with proofs represented as terms (called proof-terms), labelled formulae represented as types, and logical deduction given as type inference. The inference rules of the type theory are given in two figures: Fig 6 which gives the rules of the underlying intuitionistic logic and 7 which gives the rules connecting this logic into the full logic of proof terms for our calculus for the imperative language. We omit the full grammar of proof-terms, but the general syntax of proof-terms can be seen from the inference rules. (See [13] for full details.) Observe that, because of the ( c o n s ) rule, proof-terms corresponding to Hoare logic rule application involve proof-terms corresponding to intuitionistic logic rule application. We define a one-step proof-term normalization relation .us in Fig. 5, extending the usual /?-reduction of the lambda calculus. An application of this relation corresponds to simplifying a proof by eliminating redundant rule applications. For example, the reduction app(abstract X . C ~ - a+A~) .us , c[a/XlB
corresponds to simplifying a proof of B by deleting an occurrence of the (4-1) rule that is immediately followed by an (-+-E) inference. As in the Curry-Howard isomorphism for intuitionistic logic (see 171 or [5]) , proof normalization is given by the transitive closure of applications of this relation. Proof-term normalization does not remove any proof-terms corresponding to Hoare logic rule application. This results from the fact that a Hoare logic rule builds a program label from a given program, but there are no matching rules for reversing this construction.
Theorem 5.1. Strong normalization T h e theory LTT i s strongly normalizing: repeated application of o n a proof-term will terminate. r*g
Proof. See [13] for the proof of this and other proof-theoretic results (including the Church-Rosser property and type decidability). The theorem follows easily using rule 9 of Fig. 5, which shows that only intuitionistic 0 sub-proofs may be simplified in a Hoare logic proof. 6. Program synthesis
From a proof of a theorem, we can extract an imperative program which satisfies the theorem as a specification. For a proof/formula pair 1 A, a
356 1. 2. 3. 4. 5. 6.
7. 8.
app(abstract X . aA-'B,bA) -.+ u[b/XIB specific(use z : S. uvz:S.A,'u : S) -.+ a [ ' u / z ~ ~ [ ~ / ~ ] f s t ( ( a , b ) A A B )-.+ a A snd((a, b)AAB)
* bB
case tnI(u)AvB of Inl(zA).bc, Inr(yB 1.cC -.+ + / z ~ z case Inr(a)AVBof Inl(zA).bC, Inr(yB).cc us c[a/y] select (show(v, a)3Y.P)in zP.y.bC -.+ b[u/z][v/ylC a[b/z] and b -.+ c entails a[b/z] -.+ u[c/z] Figure 5. The eight reduction rules inductively defining -+.
program p is said to satisfy a program/formula pair 11 0 A when 0
0 0
the the the the
programs p and 1 have the same side efffects, program 1's side-effects make the formula A true, and program's return value is a return-value modified realizer for formula A (that is to say, p retmr A ) .
When this is the case, we say that p is a SML modified realizer of 1 A , and we write p mr 1 0 A. In Figs. 8 and 9 we define an extraction map extract : LTT + SML(StateRef), from the terms of the logical type theory, LTT, to programs of SML(StateRef). This map satisfies the following extraction theorems, which tell us that the map produces a SML modified realizer of the specification from a proof of that specification.
Theorem 6.1. Extraction of modified realizers from intuGiven any intuitionistic proof Flnt pA, then itionistic proofs. extract(pA) is an intuitionistic modified realizer of A . Proof. In [l,14,131, it has been shown that modified realizers can be obtained from ordinary intuitionistic proofs from the extract map. The only difference between ordinary intuitionistic proofs and formula A in the proofterm pA is that A may now involve initial and final state identifiers. However, this will not affect this proof (we can consider these identifiers to be distinguished variables in an ordinary intuitionistic proof to arrive at the 0 required result). Theorem 6.2. Extraction produces programs that satisfy proved then [extract(p)] It P . formulae. Given a proof I- pWoP,
357
A
t-lnt pAIAAz
A Elnt inr(p)Az
A t i n t pA1 A
(A-E2)
(V-11)
t i n t fSt(p)A1VA2
x is not free in A
c : T is a valid sort inference
a : T is a valid sort inference
x does not occur free in Q or A2
Figure 6. The basic rules of many-sorted intuitionistic logic, Int, presented as type inference rules, where proof-terms denote proofs and types are formulae with initial and final state identifiers. We assume that z , y are arbitrary AD7 variables of some sort T , and that a is a term of arbitrary sort T.
Proof. When P is Harrop, by the definition of extract, extract(p) = w and so [extract(p)] IF P follows from soundness (see Remark 4.1). When P is not Harrop, the proof involves showing that extract(p) always produces the same side-effect as w over all the state references that are used in P. This is routine but long and we omit it. Then the result follows from the fact that, if (u,o')IF P , then (o,o")IF P for any state u'' that differs from (TI at most on states not used in P.
358
I- Axiom(A)loA 1assign(s,
I- ite(ql,42)if
if A E AX (assign)
lzotologici (b)=false+C
I- 42
then b e l s e 111zoC
qwo(tologici(b)=trueAA[8~/B])
I- ,d(4)uhile
I>
D)s:=~osf=tologici(v)
11otologici (b)=true+C
I- 41
(Axiom -
af-then-else
1q u o A [ ~ f / ~ ]
do b;donewoA[8;/B]+A[B~/B]Atologiy(b)=false
(loop)
where A does not contain any state identifiers
Figure 7. The structural rules of our calculus. In A X , TZ is a constant, uniquely associated with the axiom A E A X .
Theorem 6.3. Extraction produces return-value realizers Take any proof I- tuoTand let L be any interpretation. Then extract(t) retmr, P.
Proof. To prove this, we take any pair of state u,u' such that extract(t) terminates with an execution sequence of the form (extract(t), u) 6 (answer, 0')
(1)
yielding a return value answer. Observe that answer has a representation answer = tologic(answer) in Term(dD7). We are then required to show that (u,a') It, Sk(P)[answer/fp]. We present only a few of the cases. Full details may be found in [13]. Case I: T is Harrop. In this case, by the definition of the Skolem form, we are required to prove that, if (w, 0)6 (answer, 0')then (0,~') It, T
(2)
359 extract ( p V o P )
any term where P is Harrop 0 uvoA
zu not H ( A )
0 H(A) fun xu => extract(a) not H ( A ) abstract uVoA.avoB extract ( a ) not H ( A ) extract(c) not H ( A ) (extract(c) extract(a)) not H ( A ) use x : T . awe* Eun x : T => extract(a) sDecificfcv*vz.A .ul -, (extract(a) v) ( P A , bW*B) (extract(a). extract(b)) case aWoAVB of inlff*A'l.bV*c. match extract(a) with I n l ( x t ) => extract(b) I Inr(x,) => extract(c) \ -
1
H(A)
V
(v, extract(a)) not H(A) (fun x => extract(b))extract(a) (fun x => fun xu => extract(b))
# ( a ) where #
f st(extract(a)) snd(extract(a)) is id, inr, fst or snd # ( w a r t f a ) ) abort( a V * l )
not H ( A )
} not H ( A )
Figure 8. Definition of extraction map extract, defined over the intuitionistic proofterms of terms used in formulae. Note that we can treat the resulting terms as MiniML program terms with state identifiers are treated as free variables.
But this is the case by soundness of the calculus (see Remark 4.1). Case 2: Proof ends in application of (loop). Assume that tnoTis of the form d,,,,
(q)while
do b;donel*A[I;/B]~(A[Bt
/B]Atologic'(b)=false)
By the induction hypothesis, we know that extract(q) retmr,
A[si/o]A tologiq(b) = true -+ A [ s f / f i ]
(3)
This means that, for any 7,T' and pure program value answer,, if (extract(q),~) 6 (answer,, 7')
we know that, for answer, = tologic(answer,), (7,~') It S k ( A [ s i / e ]A
tologiq(b) = true
-+ A[sf/ij])[answer,/fp]
(4)
where P stands for A [ s i / ~A]tologici(b) = true -+ A["/u]. There are two cases, depending on whether A is Harrop or not. We prove the more interesting latter case.
360
Then A[&/C]is not Harrop, and extract(t) is rv1 := fun x : xsort(A) => X; while b do rv2 := extract(q); rv1 := (fun xp :: x1 => fun x : xsort(A) => xp(rv1x)) rv2 r v l ; !rvl
We wish to show that answer is such that
By the definition of Skolem form and the fact that A [ % / @ and ] A [ s f / C ]are not Harrop, the required statement ( 5 ) may be rewritten
First we make some observations about the execution of the extracted program. Beginning of observations. Because we know that extract(t) terminates, by the definition of D, the program must have an execution sequence that results in states a = ao,a1,. . . ,a, = a’
where
(
[
rv1 := fun x : xsort(A) => x; while b do rv2 := extract(q); r v i := (fun x2 :: x i => fun x : xsort(A) => xz(rvix)) rvp r v l ; !rvi
(7) while b do rvz := extract(q); rv1 := (fun x2 :: XI => fun x : xsort(A) => x ~ ( r v ~ xrvz ) ) rvi;
6
(!rv1, a,) so that answer = a,(rv1), and where
36 1
and
(b, ai)D ( t r u e , ai)
(9)
(i = 1,.. . ,n - 1) and (b, %) D (false,a,)
(10)
It can be shown that (9) entails (a,,IT,+^) I!- tologiq(b) = true
(11)
for i = 1,.. . , n - 1. and (10) can be used to prove that, for any state 7 (7, a ,)
It- tologicf(b) = f a l s e
(12)
The execution sequence can also be used to show : a and ai+l differ only at rv1 and rv2 in
((
rvz := extract(q); := (fun x~ :: xi => fun x : xsort(A) => xZ(rv1x)) rv2 rvt
TVI
), ) ui
D (rvi I c i + i )
for i = 1,.. . , n - 1. By inspection of the evaluation sequence (7), ol(rv1) = f u n x : xsort(A) => x
(13)
(14)
Also, because rv1 does not occur in extract(q), the execution of extract(q) from ai to a: will not affect the value of rv1: that is, ai(rv1) = ai(rv1). So, by inspection of the evaluation sequence (8): a i + l ( r v ~ is ) the reduced normal form of fun x : xsort(A) => fll(rv2)(ai(rvt)x) for i = 1,.. . , n - 1.
362
That is, ui+l (rvl) is the reduced normal form of fun x : xsort(A) => answer,,(oi(rvl)x)
So, because answer = pn(rvl), when n normal form of
for i = 1,.. ,, n - 1.
> 1, answer must be the reduced
fun x : xsort(A) => (answer,,,-, ( f u n x : xsort(A) => answer,,,-, 14 (. . . answer,, (fun x : xsort(A) => x x) . . .)x
That is, if n
>1
answer = fun x : xsort(A) => answer,,-,(answer,,-,
. . . (answer,,x) . . .) (15)
Also, by (14), when n = 1 (that is, when u‘ = u l ) , answer = f u n x : xsort(A) => x
(16)
Take arbitrary r,r” such that (extract(q),
T)
I ;(answer,,
7“)
By the induction hypotheses (3), (4), the definition of Skolem form and the fact that A[&/ij]and A [ ~ f / i are j ] not Harrop, (7,7”)IFL
)
: xsort(A[%/ij]) .Sk(A[si/ij]) [x/f A [ s i / o ] ] Atologki(b) = true -+ A [ s f / ~ ] [ a n s w exr/,f A [ s f / c ] ]
Vx
(
(17)
Recall that rv1 and rv2 do not occur in
vx : xsort(A[Si/ij])SL(A[si/ij])[z/ f A [ s i / g ] ]A tologici(b) 4
= true
A [ s f / g ] [ a n s w e rx~/ f A [ a f / o ] ]
It can then be shown that this implies the following. For any r’ that differs from T” at most on state variables rv1 and rvg, (TJ’)
11’
(Vx
)
: xsort(A[si/ij]).SL(A[si/V])[x/f~[~~/c]] Atologici(b) = true -+ A[Sf/v][answer, “/fA[sf/o]]
End of observations. We wish to show (6): (07
a’)
b’x : xsort(A[Si/ij]) .Sk(A)[Si/V] [ x / f ~ [/ #s I;] -+
( S k ( A ) [ s f / V ] [ a n s w x/ e r f A [ s f / g ] ]A tologicr(b) = f a l s e )
(18)
363
To do this, we take an arbitrary x : xsort(A[$i/i~])-variantL‘ of assumption (a,0’)
L
with the
Sk(A)[si/vl[x/fA[s;/8]]
(19)
Sk(A)[gf/v][answer x/fA[sf/o]]
(20)
and we prove
(a,a’) and
(a,a’) It,! tologicr(b) = false
(21)
Proof of (21). By (12), ( 7 , a n ) It tologicf(b) = false for any T. So, in particular, (a0,an) lk,~ tologicr(b) = false which is the same as writing (20), as required. End of proof of (21). Proof of (20). There are two subcases: 1) a = 00 and a’ = 01 ( n = 1) and 2) G = 00 and a‘ = an (n > 1). Subcase I. In this case, by (16), answer = fun x : xsort(A) => x
in our model of ADT, and so answer x = x. It can be shown that this and (19) entail (no, a l )
Sk(A)[&/6][answer x/fA[s;/8]]
(22)
Now, observe that 00 = a and a1 = a’ differ at most on rv1, which does not occur in Sk(A)[a,/v][answer x/fA[s,/o]]. It can be shown, using the definition of It, that these facts and (22) result in (a,0‘) I ~ L ! Sk(A)[gf/fl][answer x/fA[Sf/8]]
(23)
Subcase 2. If a’ = an for n > 1, we proceed as follows. Define a1 F 2
ak
= answer,,-,
(ak)
for k = 2 , . . . ,n. As usual, we take ai to be defined as tologic(ai). It will be important to note that, as answer,, is state-free, it is the case that each ak is also state-free. Consequently, the only state references in
Sk (A)[Bf
iaj
/ f A[Sf / 8 ]1
364
are S f . By expanding the definition of a,, we obtain a, = answerun-,(answerun-,. . . (answer,,x)
. . .)
We will show, for any j = 2 , . . .,n - 1 (gj 7
aj+l)
Sk(A)[Sf/v][aj+l/fA[rf/~]]
We proceed by induction. Base case. First, note that (19) can be written as ( 0 0 , 0,)
But, because
differ only at rv1, which does not occur in Sk(A)[Si/v][z/f~[~~/~]], by reasoning about the definition of IF, we can show that this means 00
and
IF'! sk(A)[si/v][Z/fA[B;/B]]
01
((Jl,u,)
Sk(A)[si/a][z/fA[~~/~]]
Also, because final states are not used in Sk(A)[Si/o][z/f~[a~/~]], we can then derive
Sk(A)[si/v][z/fA[~;/~]l
(ul,u2)
(25)
So we can instantiate (18) with (25) and (11 with i = l),to give
( a 02) , It,! A[Sf/@][answer,, z/fA[rf/~]] and we are done. Inductive step. Assume that
,
((Jk uk+l) IF'!
Sk(A)[sf/v][ak+l /fA[Bf/B]]
holds for some k < n - 2. /0],by reasoning Because no initial state references occur in Sk(A)[Sf over the definition of IF, this means
,
( ( ~ k + l 0k+2)
Sk(A[Si/v])[~i/~I[ak+l/f~[,-;/~]I (26)
w e can instantiate (18 setting T = ok+l and r' = ak+2) With (26) and with (11)setting i = k 1) to give
+
(vk+l,uk+2) IF'!
A[Zf/v][answeru,+, ak/fA[Bj/B]]
which means
(ck+l,gk+2) 11)' A[sf/a][ak+Z/fA[rf/~)]
as required and (24) is proven.
365
Now by (24), we know in particular that (ffn-1, f f n )
Sk(A)[i?f/v][an/fA[af/o]l
Now, because initial state references do not occur in Sk(A[i?f/a]), it can be shown that this means
S k ( A )[sf/a][an/fA[iif/ B ] 1 Also, because n > 1, (15) must hold, i.e.: (go, 0,)
answer = fun x : xsort(A) => answer(,n_l,,:_l)(answer(,n_,,,~_z)
. . . (answer(,, ;,, p ). . .) we know that answer x = a,
in our model of dV‘T, and so it can be shown that (go, an)
Sk:(A)[sf/f][answer x / f A [ a /fi i ] ]
End of proof of (20). Finally, by the definition of IF, because we took an arbitrary L’ (ff,ff’) IF,
vz : xsort(A[Sz /v]) .Sk(A)[q/v] /.[ f A [ &, B ] ] + ( S k ( A ) [ g f / v ] [ a n s m e rf xA/[ g f / 0 ] A ] tologicf(b) = f a l s e )
Case: Proof ends in application of (if-then-else). Suppose that twoTis of the form b pwl otologic; (b)=true-tC 1 qw2 otologic; (b)=faZse+C (zf-then-else) ite(p,q)if then b else wlzuzoC We need to show that (u, 0’)IF, Sk(C)[answer/f c ]
(27)
Because tologici(b) = true is Harrop, so by the induction hypothesis extract(p) retmr, tologici(b) = true
+C
(28)
Similarly, tologici(b) = false is Harrop, by the induction hypothesis extract(q) retmr, tologici(b) = false
-$
C
(29)
Therefore, by definition of s and S k , (28) means that, for any states 7,T’
*
(extractb), T ) [ ;(answerp,7’) ( T ,T ’ )
IF, tologici(b) = true + Sk(C)[answer,/ f c ]
(30)
366
and (29) means that, for any states T,T'
*
(extract(q),~ ) f(answerq, j 7') ( 7 , ~ 'It, )
tologici(b) = false + Slc(C)[answer,/ fc]
(31)
Either a(b) = true or a(b) = false. We reason over these two cases to obtain (27). Subcase 1: a(b) = true. so (b, a ) D (true, a )
and so, by the interpretation of = over side-effect-free terms, this means that (a,a') IF, tologici(b) = true
(32)
Also, the operational semantics of extract(t) demands the following hold: ( i f b then extract(p) e l s e extract(q), a ) 6 (answer, a')
(extract(p),a ) 6 (answer, a')
so, (extract@),a ) 6 (answer, a')
(33)
Instantiating (30) with (33) gives (a,a') It, tologici(b) = true + Sk(C)[answer/fc]
Instantiating this with (32) gives (a,a') IF, Sk(C)[answer/fc]
which establishes (27), as required. Subcase 2: a(b) = false. Similar reasoning to the previous subcase will establish (27). Case: Proof ends in application of (cons). Suppose tWoTis of the form c o n s V P , qP+A)w*A
derived by
By the induction hypothesis, we know that extractb) retmr, P
(34)
367
There are two cases, depending on whether P is Harrop or not. We consider the more complicated, latter case. By Theorem 6.1, fun x, : xsort(P) => extract(q) is an intuitionistic modified realizer of P -+ A, and so, for any (7,~') (777')I t
vxu : xsort(P).sk(P)[xu/fP]
-h
s k ( A ) [ ( azu)/fA]
(35)
for any a = tologic(a) where a = f u n x, : xsort(P) => e x t r a c t ( q ) [ ~ ( ~ ) / ~ i ] [ ~ ' ( S ) / ~ f ]
(36)
Now, the execution of extract(t) must result in a sequence of states (T
= 00,01,02,~3 = (T'
such that
-i .-
I ;rvp := extract(p);
B
:= I;
(fn Ii :: Sf 3 fun x, : xsort(~)=> extract(q)) !I :: B !rvp
)
(fn ~i :: Sf 3 fun x, : xsort(P) => extract(q)) ,0 3 !i:: B !rvp
(37)
I ; (answer,g3) where answer = fun xv : xsort(P) => extract(q)[a3(i)/Bi]['~3(?)/Sf] 03(rvp) (38)
and
(i := s , ( T ~ )C; (a1,crl) (rvp := extract(p), 01) C; (ap,0 2 ) (P := s , ( T ~I;) ( a 2 , ( ~ 3 )
(39)
so that u1 = ao[iI+ 0,,(s)]a3= 0 2 [ I
I+ 0
2 ( ~ ) ] ~ ( r= v ap ~)
(40)
Now, because the 5 do not occur in extract(p), formula (40) and inspection of (39) reveal that 03(F)
Also, because the values of
= 01(i) = 0o(S)
(41)
are unchanged in the assignment ? := S
(Tg(3)
= (T2(I) = 0 3 ( a )
(42)
368
So, using (40), (41) and (42) in (38) gives answer = f u n x, : xsort(P) => e x t r a c t ( q ) [ a ( ~ ) / ~ i ] [ a ’ ( ~ ) / ~ ap f]
in our model of d D 7 . Define a4 = fun
xv : xsort(~)=> e x t r a c t ( q ) [ a ( g ) / ~ i ] [ d ( ~ ) / ~ f ]
So that answer = a4 ap
in our model. By (35), it is the case that (a,.’)
1 1V X U : xsort(P).Sk(P)[Xv/fp] + S k ( A ) [ ( a q%
)/f~]
Also, given that (rvp := extract(p), al) 6 (ap,uz) we let a: be the state such that (extract(p), 01) & (ap,a;) Now, recall the induction hypothesis: extract@) retmr,
P
(47)
This means that (al,.:)
IF’ Sk(P)[ap/fpI
(48)
Note that (TO can differ from a1 only on i, and a3 can differ from a: only on I and rvp. So, because i, 3 and rvp do not occur in Sk(P)[a,/fp],as for (22) ( a , 411‘ Sk(P”p/fPl
We instantiate (46) with (49) to give a’)
S k ( A ) [ ( a qap)/fA]
But then, by (45) it can be seen that
(a,a’)IFL s k ( A ) [ a n s w e r / f ~ ] as required. This last case concludes our proof.
(49)
369
7. Implementation We have implemented our calculus by encoding the LTT within SML.The proof-terms and labelled formula types are defined as data-types, the LTT typing relation is represented as pairs of terms of the respective data-types, and the rules of the calculus are treated as functions over such pairs. One of the advantages of our calculus is that it has a natural deduction presentation. This makes it easier to reason with than, say, the usual Hilbertstyle presentations of Hoare-style logics. Further, the Curry-Howard isomorphism can be exploited to enable intuitive user-defined development of proof tactics and parametrized lemmata, treated here as SML functions over proof-terms. In this way, the user can develop a proof in the way mathematicians naturally reason: using hypotheses, formulating conjectures, storing and retrieving lemmata, usually in a top-down, goal-directed fashion. The strong normalization property can also be used to simplify proofs, which is valuable in the understanding and development of large proofs.
Example 7.1. We illustrate our approach to program synthesis with an example, involving code for part of an internet banking system. We require a program that, given a user’s details, will search through a database to obtain all accounts held at the bank by the user, and then returns them in a list. For the sake of argument, we simplify our domain with the following assumptions: We assume two SML record datatypes have been defined, user and account. Instances of the former contain information to represent a user in the system, while instances of the latter represent bank accounts. We do not detail the full definitions of these types. Howver, we assume that an account record type contains a user element in the owner field, to represent the owner of the account. So the owner of the account element myAccount : account is accessed by myAccount .owner. We also assume that user is an equivalence type, so that its elements may be compared using =. We assume a constant currentUser : user that represents the current user who is the subject of the account search. The database is represented in SML as an array of accounts, db : account array
370
Following the SML API, the array is 0-indexed, with the ith element accessed by sub(db, i)
and the size of the array given as
length db Assume we have an array of size Size, accounts. Although SML arrays are mutable, for the purposes of this paper we shall consider db to be an immutable value, and therefore it will be represented in our logic as a constant. We assume a state reference, counter : int ref, to be used as a counter in searches through the database. We take a predicate
alZAccountsAt(u: user, x : account list, y : int) whose meaning is that x is a list of all accounts found to be owned by the user u,up to the point y in the database db. The predicate defined by the following axioms in AX Vu : user.Vx : (account Zist).Vy : int.aZlAccountsAt(u,x,y)-+
Vz : int.z
5 y + sub(&, z).owner = u (50)
V u : user.Vx : (account Zist).Vy : int.
< (length d b ) - l
(y
= true)Asub(db,y+l).user = uAaZZAccountsAt(u,x , y)
aZlAccountsAt(u,sub(db, y
+ 1) :: x,y + 1)
+
(51)
Vu : user.Vx : (account Zist).Vy : int. y
< (length db) - 1A isub(Z,y + l).user = u A allAccountsAt(u,x,y) + aZZAccountsAt(u,x , y Vu : user.Vy : i n k y = 0 + aZZAccountsAt(u,[I, y)
+ 1)
(52)
(53)
(these axioms are available for intuitionistic proofs, so they do not involve programs). Applications of these axioms are used in the LTT with axiom names given by their equation numbers. For instance, an application of (50) is denoted by Axiom(50).
37 1
Our requirement is to obtain a program that returns a list of accounts y : (account list) such that
EistAlIAccounts( current User,y , (length d b ) ) To extract this program from a proof, we unSkolemize this formula, to derive
3y : (account list) .listAllAccounts( currentuser, y , counterf) A
(counterf < (length db) - 1) = false
Extraction of a modified realizer for this formula will produce an imperative program whose return value is the required list of accounts. The previous axioms are Harrop. We also have a non-Harrop axiom y
< (length db)-I + sub(l,y+l).owner = u\/isub(I,y+I).owner = u (54)
Because this axiom is to be used in intuitionistic proofs, we assume that this axiom is associated with a side-effect-free program that is an intuitionistic modified realizer of (54). From (51),(52) and (54),we can derive an intuitionistic proof y
< (length db) - 1 = true, allAccountsAt(u,1, y )
+
kl,t 3 1 : (account Zist).alZAccountsAt(u,l,y 1) ( 5 5 )
By assuming 3 1 : (account list).allAccountsAt(u,l,y), we can apply 3 elimination to (55) and then obtain klntV y : int.Vu : user.(y < (length db) - 1) = true A 3 1 : (account l~st).alZAccountsAt(u, I,y)
+
3 1 : (account list).aIlAccountsAt(u,I , y
+ 1)
(56)
by (+-I) over our assumptions, and successive (V-I) over free variables. We can transform (56) into
klnt V y : int.Vw : int.w = y + 1 + (Vu : user.(y < (length db) - 1 ) = true A 3 1 : (account Zist).allAccountsAt(u,l,y) -+ 3 1 : (account list).alIAccountsAt(u,I , w)) (57)
372
We then use (58) with counteri and counterf and currentUserfor for y, v and u,to give klnt
+
counterj = counteri 1 + (counteri < (length db) - 1) = trueA 3 1 : (account list).allAccountsAt(currentUser,I , counteri) -+ 3 1 : (account list).allAccountsAt(currentUser,I , counterf) (58)
There is a proof-term corresponding to this proof, which we shall denote by and that a program PP : i n t - > int can be extracted from p5g that is a modified realizer of (58) (for brevity, we omit the full description). We also have the following, by the (assign) rule of the Hoare logic:
p58,
k counter := counter
+1
counterj = counteri + 1
This has a corresponding proof-term assign(counter, counter And so, by applying (cons) to (59) and (58), k counter := counter
+1
(59)
+ 1).
(counteri < (length db) - 1) = trueA
3 1 : (account list). alMccountsAt(current User,1 , counteri) -+
3 1 : (account list).allAccountsAt(currentUser,1, counterj) (60) The corresponding proof-term is
cons(assign(counter,counter
+ l),p5g)
Then we apply (loop) on (60) k while counter
< (length db) - 1 do counter
:= counter
+ 1;done.
3 1 : (account list).allAccountsAt(current User,1, counteri) -+
3 1 : (account l~st).allAccountsAt(currentUser, 1, counterf)A (counterj < (length db) - 1) = false (61) with resulting proof-term
wd(cons(assign(cmnter,counter
+l ) , ~ ~ ~ ) )
From the axiom (53) we can derive
counterj = 0 + 3y : (account list).allAccountsAt(currentUser,y, counterj)
(62)
with a proof-term p62. By application of (assign) k counter := O
counterf = 0
(63)
373
with proof-term assign(cmnter,0). Then, applying (cons) to (63) and (62) gives
counter := 0e3y : (account list).allAccountsAt(currentUser,y , counterf) (64) with proof-term cons(assign(counter,O ) , P ~ ~ )This . can be weakened to include a true hypothesis true:
counter := 0 true + 3y : (account list).aZlAccountsAt(currentUser,y , counterf) (65)
with a proof-term of the form cons(cons(assign(cmnter,
O), p 6 2 ) , ptrue)
where ptrue is a proof-term for an intuitionistic proof of P -+ (true + P ) ( P 3y : (account Zist).aZZAccountsAt(currentUser,y , counterf)). So, using (seq) on (65) and (61), we can obtain
t- counter := counter + I; while y < (length db) - 1 do counter := counter + 1;donee true + 3y : (account list).allAccountsAt(currentuser, y , counterf)A (counterf < (length db) - 1) = false (66) with proof-term
seq(cons(cons(assign (counter,0 ) , p62), ptrue),wd(cons(assign(counter,counter+ l),p58)))
which can be simplified to the required form I- counter := counter while y
+ 1;
< (length db) - 1 do counter := counter + 1;done.
3y : (account list).aZlAccountsdt(current User,y , counterf)A
(counterf < (length db) - 1) = false
(67)
with a corresponding proof-term of the form: cons(seq(cons(cons(assign( counter,0 ) ,
psz),ptrue),wd(cons(assign(counter, counter
+ l),p58))),qtrue)
(68)
374 p T
!xtract( t )
any term t with H ( T ) Axiomln\voA
'Kn
rv1 := fun x : xsort(A) wd(u)vhile
do b;donsl*P
where P A [ s i / e ] -+ ( A [ s f / " ]A tologicr(b) = f a l s e )
=> x;
vhile b do rvz := extract(q); rv1 := (fun x2 :: x i => fun x : xsort(A) => x2 (rv1 x)) !rv2 !rv1; !rvl: f b then extract(q1) else extract(q2) - 2
not H ( A ) , not H ( B ) and not H ( C ) : rvp := extract(p); rvq := extract(q); (fun x, :: xq => fun x : xsort(A) => rvq (rvpx)) !rvp !rvq
H ( A ) , not H ( B ) and not H ( C ) : rvp := extract(p); rvq := extract(q); rvq rvp
H ( A ) , H ( B ) and not N(C): "; rvq := extract(q); !rvq
not H ( A ) , H ( B ) , and not H ( C ) : v; rvq := extract(q); (fun xq => fun x : xsort(A) => xq x) !rvq
not H ( P ) and not H ( A ) : i := g. C V := ~ extract(p);
1 .- s;
'fn ~i :: sf => fun xv : xsort(P) => extract(q)) !rvp !I :: ? H ( P ) and not H ( A ) : I := I; v; ? := I ; (fun si :: gf => extract(q))
!I:: i Figure 9. Definition of extraction map extract from proof-terms to SML programs. Here we assume that rv1, rv2, rvp, rvq, I and 5 are state references that do not occur in extract(p) and extract(q), and whose corresponding state identifiers do not occur in any of the formulae used in the proof of p or q.
375
where qtrue is a proof of (true + A ) t A for
A
3y : (account list).aZlAccountsAt(currentUser,y, counterf) A (counterf
< (length db) - 1) = false.
Finally we apply extract to (68), obtaining the required program r v l := f u n x:account l i s t => x ; while counter<(length db) - 1 do ( r v 2 := ( i c := !counter; c o u n t e r : = c o u n t e r + l ; i f := ! c o u n t e r ; ( f u n c o u n t e r - i => f u n counter-f => (PP c o u n t e r - i currentuser)) i c i f ) r v l := f u n x-2::x-l=> f u n x => (x-2 ( r v l x ) ) ! r v 2 ! r v i ; ) ! r v l [I ;
8. Related work and conclusions Various authors have given type-theoretic treatments to imperative program logics. It has been shown in [6] how a Hoare-style logic may be embedded within the Calculus of Constructions through a monad-based interpretation of predicate transformer semantics, with an implementation in the Coq theorem prover [3]. Various forms of deductive program synthesis, with its roots in constructive logic and the Curry-Howard isomorphism, have been used successfully by [ll],[12] and [15]. The difference between our approach and those mentioned is that we do not use a meta-logical embedding of an imperative logic into a constructive type theory, but rather give a new logic that can be presented directly as a type theory. Apart from the novelty of our approach, our results are useful because they present a unified means of synthesizing imperative program according to specifications of both side-effects and return values. Further, from the perspective of theorem prover implementation, the advantage of our calculus over others is the use of a natural deduction calculus for reasoning about imperative programs and the consequent adaption of the Curry-Howard isomorphism. References 1. Ulrich Berger and Helmut Schwichtenberg.Program extraction from classical proofs. In D. Leivant, editor, Logic and Computational Complexity, International Workshop LCC '94, Indiapolis, IN, USA, October 1994, pages 77-97, 1995. 2. Robert L. Constable, Stuart F. Allen, H. M. Bromley, W. R. Cleaveland, J. F. Cremer, R. W. Harper, Douglas J. Howe, T. B. Knoblock, N. P. Mendler,
376
3.
4.
5.
6.
7. 8. 9. 10.
11.
12.
13.
14.
15.
P. Panangaden, James T. Sasaki, and Scott F. Smith. Implementing Mathematics with the Nuprl Development System. Prentice-Hall, NJ, 1986. Thierry Coquand. Metamathematical Investigations of a Calculus of Constructions. In Piergiuorgio Odifreddi, editor, Logic and Computer Science, pages 91-122. Academic Press, 1990. Patrick Cousot. Methods and logics for proving programs. In Jan Van Leeuwen, editor, Formal Models and Semantics: Volume B, pages 841-994. Elsevier and MIT Press, 1990. John Newsome Crossley and John Cedric Shepherdson. Extracting programs from proofs by an extension of the curry-howard process. In John Newsome Crossley, Jeffrey B. Remmel, Richard A. Shore, and Moss E. Sweedler, editors, Logical Methods, pages 222-288. Birkhauser, Boston, MA, 1993. J. C. Filliatre. Proof of imperative programs in type theory. In International Workshop, T Y P E S '98, Kloster Irsee, Germany, volume 1657 of Lecture Notes in Computer Science, pages 78-92. Springer-Verlag, 1998. Jean-Yves Girard, Yves Lafont, and Paul Taylor. Proofs and types. Cambridge University Press, Cambridge, 1989. Ronald Harrop. Concerning formulas of the types A 4 BVC, A + ( E z ) B ( z ) in intuitionistic formal systems. Journal of Symbolic Logic, 25:27-32, 1960. C. A. R. Hoare. An axiomatic basis for computer programming. Communications of the Association f o r Computing Machinery, 12(10):576-80, 1969. John S. Jeavons, Bolis Basit, Iman Poernomo, and John Newsome Crossley. A layered approach to extracting programs from proofs with an application in Graph Theory. In these Proceedings. Zohar Manna and Richard J. Waldinger. The deductive synthesis of imperative LISP programs. In Sixth A A A I National Conference on Artificial Intelligence, pages 155-160, 1987. Mihhail Matskin and Enn Tyugu. Strategies of Structural Synthesis of Programs. In Proceedings 12th IEEE International Conference Automated Software Engineering, pages 305-306. IEEE Computer Society, 1998. Iman Poernomo. The Curry-Howard isomorphism adapted for imperative program synthesis and reasoning. PhD thesis, Monash University, Australia, 2002. In preparation. A Technical Report, which is an extended version of this paper, is available at http: //www. csse .monash.edu. au/-ihp/karma, DSTC, Melbourne, Australia, 2001. Iman Poernomo and John Newsome Crossley. Protocols between programs and proofs. In Kung-Kiu Lau, editor, Logic Based Program Synthesis and Transformation, 10th International Workshop, L O P S T R 2000 London, UK, July 24-28, 2000, Selected Papers, volume 2042 of Lecture Notes in Computer Science, pages 18-37. Springer, 2001. Jamie Stark and Andrew Ireland. Towards automatic imperative program synthesis through proof planning. In Proceedings 13th IEEE International Conference Automated Software Engineering, pages 44-51, 1999.
377
PHASE-VALUED MODELS OF LINEAR SET THEORY
MASARU SHIRAHATA Division of Mathematics, Keio University, Hiyoshi Campus, 4-1-1 Kohoku-ku, Yokohama 223-8521, Japan E-mail: [email protected] The aim of this paper is a model-theoretic study of the linear set theory. Following the standard practice in intuitionistic and quantum set theories, we define a set to be a function from its members to non-standard truth values. In our case, the truth values are facts in a phase space as defined by Girard. We will construct the universe V p from the phase space P and verify a number of set-theoretic principles which are linear logic versions of the Z F axioms.
1. Introduction
In this paper, we will extend the Boolean-valued model for classical set theory [5,9] to linear logic. This is in analogy to the locale (Heyting)valued model for intuitionistic set theory [l],and, Takeuti and Titani’s ortholattice-valued model for quantum set theory [8]. The general idea is as follows. Given a propositional logic and its algebraic model, we can regard an element of the algebra as a (non-standard) truth value. Then we can extend the notion of characteristic functions, or sets, so that their range becomes the set of the extended truth values. In the case of linear logic, such an underlying set of truth values is given by the set of facts in a phase space as defined by Girard [3]. It is worth noting the similarity of the set of facts with the ortholattice in quantum logic. In short, the ortholattice is the lattice of closed subspaces of a Hilbert space ordered by inclusion. To each Hilbert space corresponds a physical system. Each vector in the space represents a state that a physical system can assume and each closed subspace represents an observable property of the physical system. Duals are defined by the orthogonality in the Hilbert space. Then, the correspondence is the following. 0
phase space/Hilbert space, facts/closed subspaces.
378
In fact, this is not at all surprising since Girard has this similarity in mind from the beginning: There is a Tarskian semantics for linear logic with some physical flavour: w.r.t a certain monoid of phases formulas are true in certain situations. ... One of the wild hopes that this suggests is the possibility of a direct connection with quantum mechanics ... but let's not dream too much! ([3], pp. 7-8.) The change of viewpoint occurring in linear logic is simple and radical: we have to introduce an observer, in order to verify a fact A, the observer has to do something, i.e., there are tasks p , 4,.. . which verify A ... These tasks can be seen as phases between a fact and its verification; ([3], p. 18.) The point stating the similarity explicitly here is to give the reader some assurance that the approach taken in quantum set theory can be transferred to linear set theory, at least to some extent. 2. Preliminary
In this section, we review the phase space semantics for linear logic and the construction of the Boolean-valued model.
Definition 2.1. A phase space is a quadruple P = (P,1,., I)where 0
(P,1,.) is a commutative monoid;
OIGP. 1 is the unit of the multiplication.
We will write p q for p . q. If A and B are subsets of P , we will write A . B , or AB, for the set { p q I p E A and q E B } .
Definition 2.2. Given a subset A of P , the dual or the linear negation of A, denoted A l , is defined by A' = { p E P I (Vq E A) pq E I}. Definition 2.3. A subset A of P is a fact if A = A l l . We denote the set of facts in P by Factp. Note that we always have A 5 A l l since p q E I for any p E A and q E A*. Furthermore, A G B implies B* C A' since p q E I for any p E A C B and q E B*. Then it immediately follows that A l = A l l L ,
i.e. A' is a fact for any A C_ P .
379
Definition 2.4. A fact A is valid if 1 E A. Proposition 2.1. Facts are closed under arbitrary intersection.
ni,,
We understand the arbitrary intersection Ai as the operation on the power set of P. In particular, Ai = P if I = 0.
ni,,
niE,
Proof. Let {Fi}iE, be a family of facts and A = Fi. Then, A E Fi for any i E I. Hence, A l l C F k l = Fi. Therefore, ALL & Fi. 0
ni,,
Proposition 2.2. AI" is the smallest fact containing A for any A 5 P . The order is given by the set inclusion. Proof. Let A C_ B with B = B l l . Then BL C_ A*, and this implies ALL B'-l = B. 0
Definition 2.5. We define the multiplicative operations on the set F a c t p of facts in P as follows. F @ G= ( F G ) l l , FVG = ( F L G " ) I , F -o G = ( F G l ) * . F @ G and FVG are called the multiplicative conjunction and the multiplicative disjunction, respectively. F -o G is the linear implication. FVG is defined via the DeMorgan duality as (F* @ G * ) l , and F defined as F*vG.
-o
G is
Definition 2.6. We define the additive operations on the set F a c t p as follows. F&G=FnG, F CB G = ( F u G)*% F&G and F @ G are called the additive conjunction and the additive disjunction, respectively. Definition 2.7. In addition to I,we define the constants in P as follows. 0 0
1 = I I, T=P,
I o O = T .
380
Note that I= (1)'- and T = P = 0'-. All those constants are facts since they are linear negations. Furthermore F 8 1 = F for any fact F. To see this, let p E F . 1 with p = qr for some q E F and r E 1. Then, for any s E F L , we have qs E I and qsr E 1.Hence p = qr E FLL= F . This implies F €3 1 F . The other direction follows from 1 E 1. Now we define the semantics for the multiplicative-additive fragment of linear logic (MALL). Definition 2.8. A phase structure for MALL is a phase space with a function which assigns a fact to each propositional letter. The interpretation of a sentence is a fact assigned to the sentence by extending the function inductively. Definition 2.9. A sentence is valid if the unit 1of the commutative monoid P is in its interpretation. A sentence is a linear tautology if it is valid in any phase structure. Proposition 2.3. MALL is sound and complete with respect to the validi t y in phase structure. For the proof of Proposition 2.3, we refer the reader to Girard's original paper [31. The phase semantics can be easily extended to predicate logic. We simply interpret quantifications as infinitary additive conjunction Fi and and disjunction (U Fi)'-'-.
n
Definition 2.10. Let { F i } i €be ~ a family of facts. We define the infinitary Fi and the infinitary additive sum CiEr Fi as follows. additive product
ni,,
n i E I Fi =
Fi,
~ i c Fir = (UiezFi)LL* We often omit the index set I .
ni,,
As before, we understand the infinitary intersection Fi and union UiEIFi as the operations on the power set of P. In particular, Fi = P = T and CiEr Fi = 0l'- = 0 when I = 0.
ni,,
For exponentials, we need to extend the phase space. Definition 2.11. A topolinear space is a phase space paired with the set F of the closed facts such that
381
(i) 3 is closed under arbitrary intersection (additive conjunction); (ii) F is closed under finite multiplicative disjunction; (iii) I is the smallest fact in (iv) for all A E F ,A g A = A . The linear negations of closed facts are called open facts.
Definition 2.12. We define the exponential operations on the set Factp as follows. 0 0
!F = the greatest open fact included in F , ?F = the smallest closed fact containing F .
The order is given by the set inclusion. There is a new simplified version of the definition of exponentials in the phase space [4]. For our present purpose, however, the above definition suffices.
Proposition 2.4. Linear logic is sound and complete with respect to the validity in the topolinear spaces. We now collect some useful propositions for the later calculations.
Proposition 2.5. Let F and G be facts. The following are equivalent.
IEF-oG, C G, (iii) F G ~ I. (2)
(aa) F
Proof. From (iii) t o (ii): Let FGI C 1.Then F C Gl' = G . From (ii) to (i): Let F C G. Then (1). FGL = FG' C GG' g 1.Hence 1 E ( F G l ) l = F 4 G . From (i) to (iii): Let 1 E F 4 G . Then ( l } . F G * = FGI
I.
0
Proposition 2.6. Let F and G be facts in P . Then for any p E F and q E F -oG, we havepq E G . Proof. Let r E G I . Then pr E FGI and pqr E I,since q E ( F G l ) ' . Hence pq E G I L = G. Proposition 2.7. Let A C P and F be a fact. Then, ALL @ F = ( A F ) l L .
382
Proof. AF C A l l F C (A * * F )l I = A L L 8 F . So ( A F ) * l C A l l 8 F by Proposition 2.2. For the other direction, let p E ( A F ) I . Take any q E F. Then pqr E 1for any r E A. Hence p q E A l . Since the choice of q is arbitrary, we have A"F. { p } C 1.Hence p E ( A l l F ) I . Then ( AF) ' - C_ (A"F)I and it follows that ( A L l F ) I L E ( A F ) l l . Proposition 2.8. Let Fi and G be facts. Then
(c
{(u
Proof. By Proposition 2.7, Fi) 8 G = Fi) G}lL = {U (FiG)}". Then, {U (FiG)}'l C {U (Fi 8 G ) } l l = C (Fi 8 G ) . On the other hand, Fi 8 G = (FiG)l' C {U(FiG)}l' for any i. Hence, U ( F i 8 G ) {U ( F i G ) } l l and C (Fi 8 G ) C_ {U ( F i G ) } l l by Proposition 2.2. Proposition 2.9. Let F,G and H be facts. Then
( F 8 G )8 H = ( F G H ) I L . Proof. ( F 8 G ) 8 H = ( F G ) I l 8 H = (FGH)"
by Proposition 2.7. 0
Proposition 2.10. Let {Fi} be the family of facts. Then ( n F 5 ) ' CFi.
n
=
n
Proof. It suffices to show that F? = (U Fi)'. Let p E F k and q E Fi for some i. Then pq E 1. Hence p E (U Fi)'. For the other direction, let p E (UF i ) l and q E Fi C U Fi. Then p q E 1. So p E F t . Hence p E nFk. 0 Proposition 2.11. Let F be an open fact. Then the following holds.
c
(2) F 1, (a) F 8 G G for any fact G , (ii) F = F 8 F . Proof. (i) Let F = G I with closed G. The claim follows from 1C G. (ii) F @ G C l @ G = G . (iii) Let F = G I with closed G. The claim follows from G = GvG.
Proposition 2.12. For any facts F and G, we have ! F 8 ! G = ! (F&G).
383
Proof. Note that open facts are closed under finite multiplicative conjunction. Hence ! F @ ! G is an open fact. Now !F @ !G C ! F F. Similarly ! F @ !G C G. Hence ! F @ !G C F n G = F&G. This implies !F @ ! G C ! (F&G). On the other hand, we have ! (F&G) = ! ( F & G ) @!(F&G)c_ ! F @ ! G .
s
Proposition 2.13. Let F be a fact. If F i s valid, so is ! F . Proof. Suppose 1 E F . Then F L C 1.Hence 1 = Therefore 1 E 1 C F .
ILC F * l = F . 0
We now turn our attention to Boolean-valued models. Let 23 be a complete Boolean algebra. We first define the 23-valued universe V".
Definition 2.13. We define the sets V," and the class V" by the transfinite induction on ordinals a as follows.
{
:v
= 0, Vt+l = {u I u is a function with dom(u) C V," and ran(u) = B}, Vf = V," where X is a limit ordinal.
ua.,x
v" = U a E O r d Kt3 . Ord is the class of all ordinals.
Next we define the interpretation of atomic propositions. Note that we can assign the rank p(u) to each u E Va by p(u) = t h e least a such that u E Vz+,,
Furthermore, we will use the canonical ordering [5] on Ord x Ord defined by
* [m=(a, P ) < mad?, 811 or
(a,P ) < (7,s)
[max(a,/?)= max(y,S) and a
< 71 or
[max(a,@)= max(y,S) and a = y and
P < 61.
The canonical ordering on Ord x Ord x Ord is defined similarly.
Definition 2.14. For u , v E V " , we define [u= v],[u2 w] and [uE v] by the transfinite induction on ( p ( u ) ,p ( v ) ) as follows.
*
' I[. 0
0
1') = VzEdom(v) ('(.I I[. = u ] > 7 [uC v] = AzEdom(.cl)(~(x) + [z E v]) where a .6 = = iu v] A I[ c_ . .I.
c
+ b = T a V b,
384
The idea behind the above definitions is the following translation. 0 0
*
21 E 21 (3x E w)(x = u), u g v w (Vx E u)(xE v), u=v(ju~vandvEu.
Notice that universal and existential quantifications are interpreted as infinitary conjunction (meet) and disjunction (join), respectively.
Proposition 2.14. For every u,v E V",
(i) I[ = .= 1, (ii) iu = = 1. = (iii) I[ = .A I[ =. w j 5 I[ =. wj, (iu) [uE v] A [w = u]A [t = v ] 5 [[wE t].
un,
The proof is by the transfinite induction on the canonical ordering of (p(u),p(v)). Now we extend this assignment to every sentence.
Definition 2.15. For every formula cp(z1,. . . ,xn), we define the Boolean value of cp ucp(ul, . . . , un)n
(Ul,
..., U n
E V")
as follows. (a) If cp is an atomic formula, the assignment is as we defined above; (b) if cp is a negation, conjunction, etc., u+(u1,. . . ,un)i = - w ( u l , .. . ,un)n,
111, AX(^^,. . . ,u,)n = u+(ul,. . . +,in A [x(ul, . . . ,un)n, 111, v x(ul,. . . ,un)n = w ( u l , . . . ,un)n v I [ X ( ~ .~. ,. ,un)n, 1111, + x ( u l , . . . ,u,)n = w ( u l , . . . , un)n + ux(ul,. . . ,un)n, 111, x ( u l , .. . ,un)n = 1111,+ A . . . ,un)n A XII -++ ( u l , . . . , un)n; (c) if cp is 3x11,or Vx11, 1 [ 3 ~ $ (ul, ~ , . . . ,u,n =
V A
M U , ul,.. . ,u,n,
UEV"
PTG(~, ul,. . . ,unn =
MV,
u1
. . . ,unn.
UEV"
Definition 2.16. A sentence cp is valid in V" if top element of the boolean algebra B.
[[(PI= 1, where 1 is the
Proposition 2.15. Every axiom of ZFC is valid in V".
385
3. The Phase-valued Model V p We now define our model V p . The construction is essentially the same as that of V“ except that we will use the set Factp of facts in the topolinear space (P,F)instead of the boolean algebra B. Definition 3.1. We define V,” and V p by the transfinite induction on ordinals a as follows:
{
vop = 0,
Vz+l = {u I u is a function with dom(u) C V’, and ran(u) = Factp}, V r = Ua<XV,” where X is a limit ordinal.
v p = UaEOrd v,” *
The rank p(u) is defined similarly as in V”.
Proposition 3.1. VF C V,” for
0 < a.
Proof. The proof is by the transfinite induction on a. Assume that V: C V g holds for any y < p with 0 < a. If a is a limit ordinal, the proposition holds by the definition. Suppose that Q = a’ + 1. Let u E V r with 0 < a, and p‘ = p(u). Then /?‘+ 1 5 p and u E Hence dom(u) VF with p’ < a‘. So dom(u) C V$ by the inductive hypothesis. Hence u 6 V z . 0 Definition 3.2. For u , v E V p , we define [u= v],[u & v] and [uE v] by the transfinite induction on ( p ( u ) ,p ( v ) ) as follows.
.[I E .]I = C z E d o m ( v ) ( v ( z ) 8 8. = 4l) , I[. C ‘1 = nz,,om(u,(u(.) --o.[I E vll), .1 = = !I[. 2 8 !I[. E .I. We note that .[I E .]I
= 0 when dom(v) =
0, and [uC .]I
= T when
dom(u) = 0.
Proposition 3.2. For every u,v E V p ,
(i) 1 E u(z) -O [z E u]for all z E dom(u), (ii) 1 E I[ =. u], (iii) 1 E = = .I.
nu
Proof. We prove (i) and (ii) by the simultaneous transfinite induction on p(u). Note that the base case p(u) = 0 is subsumed under the case dom(u) = 0.
386
(i) We show that for all z E dom(u),
I[ E.
I.
If dom(u) = 8, this holds trivially. Otherwise, 1 E [z = x] by the inductive hypothesis. Then u ( z ) = u ( z ) . (1) (u(z) . . I[ E. unL = IL c - UyEdom(u)(u(Y) . uz = ~ence 4.1 . (UyEdom(u) ( 4 ~ 8b ) = Yll))"' = ' (Uy€dom(u) ( u ( Y ) @ I[ =. 1. (ii) It suffices t o show 1 E !I[. u].NOW [U u]= --o Otherwise, [z E u]). If dom(u) = 0, we have 1 C P = [uC u]. u ( z ) . [z E .]IL I for all z E dom(u) by (i) so that 1 = n,Edo,(u,(u(z) --o [z E u]).Since Iis the smallest closed fact, 1 is the greatest open fact. Hence 1 = ![uC u]and clearly 1 E 1. (iii) We show [u = w] . [v = u ] ' C 1.Since the multiplicative conjunction is commutative, [u = v] = [v = u ] and the claim immediately follows. 0
w*.
w) mLc
c
nzEdom(u)(+)
c
c
c
Proof. The proof is by the simultaneous induction on (p(u),p(v),p(w)) along the canonical ordering, which is sensitive to the permutation of u,w and w. To carry the induction through, we divide (ii) and (iii) into two separate cases:
(ii-a) (ii-b) (iii-a) (iii-b)
1E 1E 1E 1E
lv E wi 8 I[ = .--o I[ E wi, . uv E 8 I[=.w]-+ E ui, uu E 8 I[=. --o I[ E wi, . 8
E
nu
=
E
So, we in fact prove (i), (ii-a), (ii-b), (iii-a) and (iii-b) simultaneously.
c
(i) First we show [uC v] 8 [w = w] [uC 2.1. By Proposition 2.2, Definition 2.5 and Definition 3.2, it suffices to show
n
).(.(
zEdom(u)
--o
8.
E
n
(W
z Edom( u )
I[ E. wn).
387
Let p E n,E,o,~,,(u(z) --o [z E u])and q E [v = w].We want to show p q E ( u ( y ) [y E w]l)lfor any y E dom(u). Let T E u ( y ) . [y E w ]' and T = tt' with t E u(y) and t' E [y E w]'. Since p E u(y) 4 [y E w],we have pt E [y E w].By the inductive hypothesis of (iii-a),
Hence ptq 6 [y E w].Therefore pqr = ptqt' E I,and it is done. Similarly, we can show [w 5 v] 8 [u= v] 5 [w C u] with the roles of u and w being exchanged, using the inductive hypothesis of (iii-b). With the above preparation, we show [u= v] 8 [v = w] [u= w].Using Proposition 2.11,
c
Now open facts are closed under finite multiplicative conjunction. Hence ![uC_ w] 8 [w = is open. Therefore, we have ![ug v] 8 [w = u] ![u5 w]. Similarly, ![w w] 8 [u= u] 5 ![.I u]. Hence, [u= w] 8 [U = w]C [u= w]. (ii-a) We want to show [u E w]8 [u= w] C [u E w].By Proposition 2.8 and Proposition 2.9,
c
E w]8
c
= .] =
(CC I €do,(
)
( ' ~ ( 8~ I 1[ =. .])
I Edom(w)
=
c
8 [U = .]
(w(z) . I[ =. .]I. I[ =. .]) 'I
w)
By the inductive hypothesis for (i),
Hence [V E w ] 8 [U = .] E &dom(w)(~(z) 8 [X = u ] )= [U E w]. (ii-b) is proved similarly, using the inductive hypothesis of (i). Note that (i) is symmetric with respect to the roles of u and w.
388
c
(iii) We want t o show [uE u]8 [[w = w] [uE w]. By Proposition 2.8, Proposition 2.7 and Proposition 2.9, E .] 8 I[ =. w] =
(
).(.(
8
z€dom(v)
C
=
C
=
qw C_
. !I[.
).(
I[.1
C_ w ] .
= w]
8
II
I[. = .])
zEdorn(v)
.K. c n
Let y E dom(v). By Proposition 2.6,
m
c s V(Y)
![. wn
wn
= w(y) . (
).(w
--o
u.
E wn)
x€dom(v)
c v(y) . MY) c uy E wn.
--o
BY E wn)
Therefore it suffices to show [y E w j .
uy = c .1 E w].
c
However we have [y E w] 8 [[y = u] [uE w] by the inductive hypothesis of (ii-a). Hence, by Proposition 2.11, ~[yE w]
.
c
.I [ = ~
g
E w] 8 qw
c I[Y E wn 8
c
8
uy =
= un
2 I[ E w]. . (iii-b) is proved similarly with the roles of u and w being exchanged, using the inductive hypothesis of (ii-a). 0 Definition 3.3. For every formula cp(z1,. . . , xn), we define the phase value of cp
[v(u1, . . . ,un)n
(u1,. .. ,u,E V P )
as follows. (a) If cp is an atomic formula, the assignment is as we defined above; (b) if cp is a linear negation, multiplicative conjunction, etc.,
w ( u l , . .. = [+(ul,.. . ,un)nL, 8 x(ul,.. . ,unn = u+(ul,. . .,un)n8 [X(ul,. . .,un)n, I [ + V X ( ~ ~. ., . ,unn = w ( u l , . . . , u n ) m ( u l , . . . ,un)n7
389
Proposition 3.4. 1 E [u=
@ I[+(.)]
Q
[+(v)Jj for any formula
4.
Proof. The proof is by induction on the construction of 4, using that [u= v] is an open fact. For example, 1 E [u= u]8 [$(v)] Q [+(u)]implies 1 E I[. = W] @ I[$+)] 4 [$I(v)]. Similarly 1 E .[I = v] @ [$(u)] 4 I[$(u)ll and 1 E [U = .]I @ [X(u)B--o [x(w)] implies 1 E [u= u] @ [Q @ x(u)l 4 @Xwn. 0
u+
We now start checking the validity of the basic set-theoretical principles. We will write (3y E z) +(y) and (Vy E z) +(y) for 3y (y E IC 8 $(y)) and Vy (y E z --o d(y)), respectively.
Proof.
(a) By Proposition 2.8 and Proposition 3.4,
UP^ E +wn
=
C (UYE n. 8 umn) yEVP
390
yEVP zEdom(s)
z Edom( z)
For the other direction, note that if y E dom(z), we have z(y) [y E z] since 1 E [y = y]. Hence,
C
MY)
€3
u$(dn)c_
y €don( z)
C ( l [ ~E zn €3 [wn) C ([Y E zn 8 nwn)
yEdom(z)
E
gEVP
= u3y (Y E
2
8 4wn.
(b) The proof is by Proposition 2.10 and (a).
0
Let us verify a number of formulas which are intended to be the linear logic counterparts of the classical ZF axioms. For the moment, we do not have any canonical way to translate classical set-theoretical formulas into linear logic. In particular the choice between multiplicative and additive connectives are rather arbitrary. Theorem 3.1. The following formulas are valid in V p .
(Empty Set):
3YVz(zE Y)l.
(Extensionality):
VXVY(!Vu(u E x -0u
E
Y)€3!Vu(uE Y -Ou E X ) -0 x = Y).
(Pair): VuVv3aVz(z = u @ z = o -0 3: E a ) . (Union):
VX3YVu(3z(z E X 8 u E z) 4 u E Y ) .
(Separation):
VX3Y(!Vu(u E Y -0 u E x 8 $(u)) 8 !Vu(u E x €3 $(u)-0 u E Y ) ) . (Collection): Vu((Vz E u)3yd(z,y)
-o
3o(Vz E u)(3y E w ) +(z,y)).
391
(Infinity): 3 Y ( ! 0 p E Y €3!Vx(x E Y -ozu{x} E Y ) ) .
0 p and X U{x} are the elements of V p which will be explained in the proof. Proof. (Empty Set): Let Y E V p be such that dom(Y) = 0. Then, for any x E V p ,we have UvEdom(Y)(Y(w) €3 [[x= w]) = 0. Then
(U
( Y ( v )8 [X =.]I)
vEdom( Y )
Hence, [(x E Y)'] = P and p x ( x E Y ) q = P. Obviously, 1 E P. (Extensionality): Recall f X = Y ] = ! [ X g Y ] ~9 ![Y g X ] . Then the axiom holds by Proposition 3.5.
(Pair): Let a E V p be such that dom(a) = {u,w} with 1 E a(.) and 1 E a(.). Then, 1 E [uE a] and 1 E [w E a ] . Now for any x E V p , we have [x = u]€3 [uE a] g [x E a] so that [x = u] [x E a]. Similarly, [x = u ] c [X E a ] . Hence [X = u @ x = u] = ([x = u]u [X = V])ll E
.]**
=
E
.I.
(Union): Let Y E V p be such that dom(Y) = U { d o m ( z )I z E dom(X)} and for any u E d o m ( Y ) , writing S(u) for { z I u E dom(z) and z E dom(X)},
Y ( u )=
c
( X ( Z >8 z(.)>.
Z€S(U)
Then for any z E dom(X) and u E dom(z), we have X ( z ) €3 z ( u ) C Y ( u ) C [uE Y ] so that X ( z ) C z ( u ) --o [uE Y ] . Hence, by Proposition 3.5, X ( z ) g [Vu(u E z --ouE Y ) ] . This means that 1 E p z ( z E X 4 Vu(u E z --o u E Y ) ) ] .The axiom then follows.
(Separation): Let Y E V p be such that dom(Y) = dom(X) and Y ( u ) = X ( U )€3 [p(u)] for all u E dom(Y). Then Y ( u ) C [uE X €3 p(u)] for all u E dom(Y) so that 1 E [Vu(u E Y -ou E X €3p(u))] by Proposition 3.5. Also X ( u ) €3 [p(u)] = Y ( u ) C [uE Y ] for all u E dom(X) so that X ( u ) g [p(u)] -O [u E Y ] . By proposition 3.5, 1 E E y ) ) ]= pu(. E x €3 p(u) u E y)n. p q u E x --o (p(u)
+,
392
(Collection): Given z E Vp, let
Fz = {S E Factp 139 E Vp([cp(z,y)] = s)} Then F, is a set and (Vs E Fz)3a3y ([cp(z,y)] = s and p ( y ) = a ) . Hence by the Collection principle of ZF, 3 4 V s E F X ) ( 3 aE v)3y ([cp(z,y)] = s and p ( y ) = a ) .
Let
CY, =
U{CY+ 1 I
Q
E Y and
LY E Ord}.
Then
393
(Infinity): We denote the phase-valued set obtained in (Empty Set) by {z, {z}} and z U {z} denote the phase-valued sets obtained by (Pair) and (Union) for now. Define Y E V p in such a way that
0 p . Similarly
dom(Y);
0
0p E
0
if z E d o m ( Y ) , then z U {z} E d o m ( Y ) ; 1E Y(0p); Y ( z )g Y ( zU {z}) for all 2 E dom(Y).
Notice that if p ( z ) = a , then p ( { z } ) = a + l and p ( { z , { x } } ) = a+2. Also, p(z U {z}) = p ( U { z , {z}}) 5 p ( { z , {z}}). Hence, given 0.p E V,”, we can have Y E V,”++w+l.Therefore, Y E V p . Since 1 E Y(0.p)5 [ 0 p E Y ] ,it suffices to show
Now for any z E d o m ( Y ) ,
4. Relating
V F to the Heyting-valued model
Let’s begin with the following observation, which is what is behind the Girard’s second translation [3] of intuitionistic predicate logic into linear predicate logic given as follows.
A* = ! A for A atomic, ( A V B)* = A* @ B*, ( AA B)* = A* 8 B*, ( A 3 B)* = ! (A* 4 B * ) , O* = 0, @A)* = 3zA*. (VzA)* = !Vz A*, Proposition 4.1. Let 0 be the set of all open facts in P . Then 0 is a frame (locale or Heyting algebra) in the sense of Vickers [lo].
394
Proof. The order is given by the set-inclusion. The arbitrary join V Fi is defined as (U F i ) l l . Let {Fi} be a family of open facts. By Propowhere F: are closed facts, so that sition 2.10, (UFi)l* = (UF i ) l l is an open fact since closed facts are closed under arbitrary intersection. The binary meet of the open facts F and G is F @ G, since F @ G = ! F @ ! G = ! ( F n G) by Proposition 2.12. Furthermore the distributivity of the binary meet over the arbitrary join holds by Proposition 2.8. 0 Frames and Heyting algebras are the same structures, although they differ when we consider homomorphisms. Locales are frames together with the set of "points." For our purpose, it is harmless to use those terms interchangeably. Then we can obtain the Heyting-valued universe V o as the subuniverse of V p by restricting facts to open facts in the construction of V p . Definition 4.1. We define V," and V o by the transfinite induction on ordinals a as follows.
v,0=0, 0
0
V s , = {u I u is a function with dom(u) E V," and ran(u) = O } , V p = Ua<XV," where X is a limit ordinal, = Ua€OrdVcuO.
'v
Definition 4.2. A phase-valued set u E V p is static if u E V o . Proposition 4.2. Let all x E v P .
E
V p be static. T h e n [x E u] is a n open fact for
Proof. [x E u]= CzEdom(u)(u(z) @ [x = z ] ) . Since open facts are closed under finite multiplicative conjunction, u ( z ) @ I[x = z ] is an open fact. Furthermore, open facts are closed under infinitary additive disjunction by Proposition 2.10. Hence [x E u]is an open fact. I3
We introduce the restricted quantifications over Vo as follows.
w $ ( ~ ,ul, . . . ,u,)n = I[v.*$(~, ul,.. . ,un)n =
zvEVo I[$(., ul,. . . ,un)n,
nVEvo I[+(., ul,.. . ,un)n.
Proposition 3.5 holds with those restricted quantifiers as well. The proofs are exactly the same.
395
For the counterparts of the power set axiom and H. Friedman’s Einduction [2,6], it seems that we need to use those restricted quantifiers.
Theorem 4.1. The following formulas are valid in V p .
(Static Set): Vx*Vy (y E z -o ! (y E z)). (Static Power Set): Vu*3v*Vz*(!V’y(y E z -o y E u)-o x E w)). (Static €-induction): !Vz*((Vy (y E z -o $(y))
-o $(z)) -o Vz*$(z).
Proof.
(Static Set): This follows from Proposition 4.2. (Static Power Set): Let w E V o be such that dom(w) = {f
I f is static with dom(f)
= dom(u)}
and 1 E w(x) for all x E dom(v). We want to show 1 E pz*(!Vy(y E z -oy E u)-o x E w)]. For this, we define z’ E dom(v) for each x E V o which satisfies
I[!vy(y E
E
-o
= qz
c c iz‘ = .I.
Given such an z’, the validity of the formula immediately follows since 1 E [x’ E w] and !I[. c_ [x‘ = x] [z‘= z] 8 [z’ E .]I [x E w] for all x E vo. The definition is as follows. Given z E V o , let x’ E V o be such that dom(z’) = dom(u) and z‘(y) = [y E z] for all y E dom(z‘). Clearly, 2’ E dom(v). Now for any y E V p ,
c
c
c
C C
[Y E x’n =
( ~ ‘ ( 2 )8
I[ =. Yn)
zEdom(u)
=
(UZ
E zn 8 I[ =. yn)
r Edom( u)
c UY E 4. Hence 1 E [Vy(y E x’ -oy E z)] = [x‘
ly E
8 E
=
c z]. Next for any y E V”,
C w)8 uz = yn 8 b E .]I) zEdom(u)
E
C zEdom(u)
( [ z = y] 8 I[ E. since u ( z ) is open
396
z€dom(u)
They are special consequences of the more general principle. Proposition 4.3. For the static u and v, our definitions of [uE w] and [u= v] in V p yield the same open facts as the Heyting-valued interpretations an vo.
Proof. Since the meet in 0 is the tensor in P and the supremum coincides in both of them, it suffices to confirm that !I[. C v] in the phase-valued model is the same as [u v] in the Heyting-valued model for u , v E V*.
c
397
n
Note that the infimumof open facts Fi in 0 is given by ! Fi. Furthermore ! Gi = ! !Gi holds for any family {Gi}iElof facts in ?. To see this, just notice ! Gi E !Gi for any i E I. Hence we only need to show that !(u(z) -O [z E u]) in P is indeed u(z) 1%E u]in 0. Now F@!(F-G) C G holds for any facts F and G. Suppose F @ H 5 G for open facts F , G and H . Then H C F 4 G and H ! ( F4 G ) since H is open. By the uniqueness of the residual F G, we can conclude that
ni,,
n,,,
n
+
+
! ( F +J G ) = ( F + G).
Since u ( z ) and [z E u]are open facts, the above argument shows that !(u(z)4 [z E u])is the same as u(z) [z E u]. 0
+
Then the formulas in the intuitionistic set theory evaluated in V o retain
the same interpretations under the Girard’s second translation with all the quantifiers modified t o the restricted ones. Furthermore, if the quantifiers are bounded, then there is no need to restrict them due t o Proposition 3.5. We hope to explore this point in more detail in the sequel of this paper. Acknowledgments The early version of this paper is included in my dissertation “71. I thank the anonymous referee for the meticulous and helpful comments.
References 1. M.P. Fourman and D. Scott. “Sheaves and logic.” Application of Sheaves, Springer Lecture Notes in Mathematics 753, 1979, 302-401. 2. H. Riedman. “The consistency of classical set theory relative to a set theory with intuitionistic logic.” The Journal of Symbolic Logic 38, 1973, 315-319. 3. J.Y. Girard. “Linear logic.” Theoretical Computer Science 50, 1987, 2-102. 4. J.Y. Girard. “Linear logic: Its Syntax and Semantics.” Advances in Linear Logic, (eds.) Girard, Lafont, Regnier, London Mathematical Society Lecture Notes Series 222, Cambridge University Press 1995, 1-42. 5. T.J. Jech. Set Theory, Academic Press, New York, 1978. 6 . A. SEedrov. “Intuitionistic set theory.” Harvey Friedman’s Research on the Foundations of Mathematics, Studies in Logic and the Foundations of Mathematics Vol. 117, North-Holland, 1985, 257-284. 7. M. Shirahata. Linear Set Theory, dissertation, Department of Philosophy, Stanford University, 1994. 8. G. Takeuti. ”Quantum set theory.” Current Issues in Quantum Logic, Plenum, New York, 1981, 303-322. 9. G. Takeuti and W.M. Zaring. Asiomatic Set Theory, Springer, New York, 1973. 10. S. Vickers. Topology Via Logic, Cambridge University Press, 1989.
398
A PROBLEM ON THEORIES WITH A FINITE NUMBER OF COUNTABLE MODELS
AKITO TSUBOI Institute of Mathematics, University of Tsukuba, Ibaraki 305-8571, Japan E-mail: [email protected] In this paper we discuss non-w-categorical theories with a finite number of countable models. The question remains open whether such theories should be unstable or not. We briefly explain known results concerning this question, and discuss related topics.
1. Introduction In this paper we discuss the following problem and related topics.
(*) Is there any non-w-categorical stable theory T with a finite number of countable models? Nowadays it is widely believed that there exists such a complete theory. However, in the author’s opinion there are not so many works aiming for constructing such theories. Many authors have proved the non-existence of such theories under additional finiteness conditions. (See [8],[3],[5],[12]and [19].) Such finiteness conditions are, for example, superstability, admitting finite coding, supersimplicity and so on. Superstability is a finiteness condition on the length of forking sequences of types. Admitting finite coding is a finiteness condition on the size of canonical bases. Unfortunately, without these assumptions, very few things are known. Since the instability is equivalent to the order property (existence of order structures in a very weak sense), (*) can rewritten as:
(**) Does any theory T with a finite number (> 1) of countable models have the order property? In fact, in [12],[19]and [4],they found definable order structures in theories with finitely many countable models (under some additional finiteness
399
conditions). In section 1, we recall some examples of non-w-categorical theory with a finite number of countable models. We also recall some basic facts for these theories. For the reader’s convenience, outline proofs will be given. In section 2, we consider theories with three countable models. By Vaught’s result,. non-w-categorical theories have at least three countable models. If a given theory is not small, of course it has infinitely many models. If T is small, there are at least three countable models: a prime (and atomic) model over 0, a prime (and atomic) model over a finite tuple realizing a non-principal type, and a countably saturated model. In [4],by assuming the almost w-categoricity of T , the existence of dense linear order in T was shown. A slight generalization of this fact will be given. In section 3, we discuss Lachlan’s theorem ([8]) and it’s generalizations. The theorem states that there is no superstable theory T with 1 < I ( w ,T ) < w . Besides Lachlan’s original proof, there are several other proofs (see [9],[14] and [15]). In these newer proofs, in particular in [9], the open mapping theorem (or a weaker version of it) seems to have an important role. In a simple unstable theory, the open mapping theorem does not hold in general. But Kim [5] showed that Lachlan’s theorem can be generalized t o supersimple theories. In (201, a restricted version of the open mapping theorem was shown for simple theories. We give a proof of Kim’s result using this restricted version. In this paper, T denotes a complete theory formulated in a countable language. I ( w , T ) denotes the number of nonisomorphic countable models of T . Models of T are denoted by M ,N , .... We fix a sufficiently saturated model of T . Usually we work in this model. A , B , .. will denote subsets of this model. Finite tuples of elements in this model are denoted by a , b, .... We write AB for denoting the set A U B. Finite tuples of variable are denoted by x,y, .... Types are complete types unless otherwise stated. S(A) denotes the set of all types over A. If p is a type, p M denotes the set of all realizations of p in M .
2. Examples and Basic Facts. As is well-known, w-categorical theories T are characterized by the property that for each finite tuple x of variables, T has only finitely many types in x. Such a good characterization is not known for theories with 1 < I ( w ,T ) < w . First we recall basic facts.
400
Fact 2.1. (1) Let I(w,T) 5 w. Then T is small, i.e. S(0) is countable. So f o r each finite set A there is a prime model over A. There is also a countably saturated model of T . (2) (Vaught) IfI(w,T) > 1 then I(w,T) 2 3.
Proof. (1) Each type in S(@)is realized in some countable model. On the other hand, each countable model realizes only countably many types. So we have (S(0)(I II(w, T ) ( w. (2) We can assume 1 < I(w,T) 5 w. So there are a prime model MO and a countably saturated model M2. Let a realize a non-principal type and let M I be a prime model over a. Then easily we have M I Mo, M z o
+
The most well-known and fundamental example with three countable models is the following one due to Ehrenfeucht.
Example 2.1. Let T = Th(Q,<, 0,1, ...). Then I ( w , T ) = 3. Other than the standard model (Q,< 0,1, ...), there are two nonstandard models M I and M2. M I has the nonstandard part { u E M I : a > 0 , a > 1,...} with the minimum element, while the nonstandard part of MZ has no minimum element.
If we want to have a theory T with I(w,T) = n(> 3), we only have to prepare n - 3 new unary predicates U1,...,Un--9 with the properties (i) the interpretations of these predicates give a partition of (Q,<, 0,1, ...) and (ii) each predicate is a dense subset of (Q, <, 0,1, ...). (See [2].) There is also an example with a dense tree structure.
Example 2.2. Let M = (M, < , A ) be a dense w-branching tree with the meet operator A. Let {ci : i < w} c M be a strictly increasing sequence. Then the theory of (M, <, Q, c1, ...) has three countable models. Miller [lo] gave a series of examples using such tree structures. The two examples above are constructed from unstable w-categorical theories by adding infinitely many constants. By the following fact we see that we cannot construct a non-w-categorical stable theory with a finite number of countable models using a similar construction to that mentioned above.
Fact 2.2. ([19]) Let T be a complete theory with 1 < I(w,T) < w. If any reduct of T to a finite language is w-categorical, then T is unstable, in fact it has the strict order property.
401
All the examples given above have infinite languages. However there is a complete theory T in a finite language with 1 < I ( w , T ) < w . Example 2.3. Let L = {+}, where 4 is a binary predicate symbol. Let T be the theory of the following structure.
(1) 4 is a preorder of the universe. The associated ordered set is isomorphic to the rational numbers Q. (2) For each n, let X , be the set of all classes having 5 n elements. Each X , has the maximum element (class), and {X, : n < w } is a strictly increasing sequence of initial segments of the preordered set.
For the same reason as in Example 2.1, we have I ( w , T ) = 3. Even though this example has a finite language, it @-interpretsthe theory of Example 2.1. Concerning extensions of a theory by constants, there are several results. Fact 2.3.
( I ) (Millar [lo]) There is a theory T with infinitely m a n y countable models such that some extension of T by a constant has only finitely m a n y countable models. (2) (Peretyat’kin [ll]) There is a theory T with three countable models such that any extension of T by a constant has infinitely m a n y countable models. According to Pillay [14], we will say that a type q(x) = tp(a/b) is semiisolated if there is a formula p(x, b) E q(z) such that p(x,b) -/ tp(a). In [13], he says a semi-isolates b if tp(a/b) is semi-isolated in the above sense. Notice that the relation on tuples given by semi-isolation is a preorder. Fact 2.4.
(1) (Benda [l])If 1 < I ( w , T ) < w , then there is non-principal type p ( x ) E S(0) such that whenever a model realizes p then it realizes all types in S(0). Such a type p is called powerful. (2) (Pillay [14]) Let p ( x ) be a powerful type. Then we can find two realizations a and b such that (i) tp(a/b) is semi-isolated and (ii) tp(b/a) is not semi-isolated.
402
Proof. (1) Suppose that there is no powerful type. Then we can inductively choose an increasing sequence of complete types {pi(xi)}iEu and a sequence of countable models {Mi}iEwsuch that Mi realizes pl, . . . , p i but does not realize pi+l. Then Mi’s are mutually non-isomorphic. (2) Let a1 be any realization of p . By compactness, we can choose a2 p such that tp(a2lal) is not semi-isolated. Let a3 be another realization of p and M a prime model over a3. Since p is powerful, we may assume that a1a2 C M . Then tp(aila3) is semi-isolated, whereas tp(aa/al) is not semi-isolated. 0 3. Theories with three countable models.
A model is called weakly saturated if it realizes all types in S(0). If I ( w , T ) = 3, then the three countable models are a prime model, a weakly saturated (non-saturated) countable model and a countably saturated model. (See [all.) Theories with exactly three countable models have strong properties, which are not expected for those with more than three countable models (see facts below). In this section, we assume I ( w , T ) = 3 unless otherwise stated. The following two facts are easy to verify.
Fact 3.1. Every non-principal type is powerful. In other words, non-prime models are weakly saturated. Fact 3.2. L e t p ( x ) E S(0) be a non-principal type and M a model. If f o r any realization a E p M there is a realization b E p M such that tp(b/a) is n o t (semi-)isolated, then M i s saturated. Definition 3.1. We will say that T is almost w-categorical if for any types pl(xl), . . . , p n (xn) E S(0), pl(x1) U...Up,(x,) has afinite number of completions. All known theories with a finite number of countable models interpret a dense linear order. We don’t know whether there is a theory T with 1 < I ( w , T ) < w such that T does not interpret a dense linear order. However, with the assumption of the almost w-categoricity, we know the following:
Fact 3.3. ([4]) If a theory T with three countable models i s almost w categorical, then T interprets a n infinite dense linear order using parameters. Here we prove a slightly stronger result:
403
(*) Let T be an almost w-categorical theory with 1 < I ( w , T ) < w . If every weakly saturated (and non-saturated) model is prime over a finite set, then T interprets an infinite dense linear order using parameters (cf. Fact 3.2). Proof. As in [4], the denseness follows easily. So we prove the existence of an infinite linear order. Let M I ,..., M , be the list of all weakly saturated non-saturated models modulo isomorphisms. For each i, choose a tuple ai of elements in Mi such that Mi is prime over ai. We may assume that all ai's have the same length. Notice that p i = tp(ai) is a powerful type. Let M be a countably saturated model and D the set Uii,p?. By the almost w-categoricity and the fact that T is small, there are formulas pi($) E pi(z) (i 5 n) and a binary formula z 5 1~ such that
(1) a 5 b w tp(b/a) is semi-isolated, for any a, b E D; (2) p i ( z ) is the unique non-principal type extending pi(z) (i 5 n), In general 5 is a preorder on a definable set extending D. Modulo the equivalence relation 2 5 y 5 z, we have the associated order. For simplicity, we may assume that 5 is an order from the first. Let a E D. We show that the set defined by {z 5 a} U p ( z ) is linearly ordered by 5. Assume otherwise. Then there are two elements b, d E D with (i)b, d 5 a, (ii)b $ d and (iii)d f b. So in a prime model over ad, we can find b' 5 a with b' $ d and d b'. But b' does not realize any pi(x), since otherwise we would have b' 2 d. So, by the condition (2) above, we can assume that b' realizes a principal type. Since each pi is a powerful type, we can choose e having the same type as b' such that e <_ D, i.e. e <_ z for any z E D. For each i 5 n, let I'i(z) be the following set of formulas: pi($) U {z
zf a : a E D}.
Since tp(e) is an isolated type, I'i(z) cannot be implied by a single formula over {e}UD. (Use the almost w-categoricity.) So there is a countable model M* 3 { e } U D that omits each I'i(z). So e 5 u p ? ' . M * is weakly saturated, since it realizes powerful types p l , ...,p,. Moreover, each p?' has no minimal element with respect to 5. So M* cannot be isomorphic to neither of the Mi's, hence M* is a saturated model by our assumption. Since M* is saturated, there must be an element d' E M * with tp(ed) = tp(ed'). Then we have that e and d' are not comparable. This is a contradiction. 0
If an L-theory T with I ( w , T ) = 3 is almost w-categorical, then any extension of T by finitely many constants has finitely many countable models.
404
This can be seen as follows. Let T, be any complete extension of T by the constants c and p ( z ) the L-type of c. ( p ( z ) is assumed t o be a powerful type.) By the almost w-categoricity, there are only finitely many (say n ) completions of q(s) U q ( y ) . By way of contradiction, assume that there are non-isomorphic countable models M I , ...,Mn+2 of T,. If the L-reducts MilL and MjIL are saturated, then they are isomorphic as T,-models. So we may assume that MIIL,...,MnIL are all weakly saturated. So there are di (i = 1,..., n) such that each MilL is atomic over di. Indeed, each MilL is atomic over cM;di. By our choice of n, there are i # j with tp(cMidi)= tp(cMjdj). Then Mi and Mj are isomorphic, a contradiction. So the theory constructed in [ll] is an example of non almost wcategorical theory with three countable models. However the theory constructed in [ll]interprets a dense binary tree, so it interprets a dense linear order (using parameters). 4. Stability, Simplicity and Countable Models.
In early 70’s Lachlan [8] showed that there is no superstable theory T with 1 < I ( w , T ) < w. His proof uses his Rank, which assigns an element of ww to each formula. Besides Lachlan’s original proof, there are several other proofs. In [9] Lascar proved a version of the open mapping theorem for superstable theories. (The open mapping theorem will be explained below.) Using it, he gave a very nice proof of Lachlan’s result. Shelah introduced the notion of weight, and gave another proof of Lachlan’s proof (see [IS]). Pillay introduced the notion of semi-isolation and gave a very short proof (see [14]). The proofs in [16] and [14] do not use the full strength of superstability. In fact, the following condition (FW) (a finiteness condition for weight) is sufficient. Since the open mapping theorem holds for any stable theory, Lascar’s proof can also be done under (FW). (FW) Let I be an A-independent set with a L Ab for all b E I. Then I is finite. Lachlan’s result stated above has been strengthened in many ways. In the following, we explain some such results. In [3] Hrushovski showed that if T is small (i.e. IS(@)lis countable) and T admits finite coding (i.e. every global type is based on a finite set) then any type has a finite weight in the sense of (FW) above.
Fact 4.1. ([3]) Let T be a stable theory such that every global type is based o n a finite set. T h e n T is w-categorical, or T has infinitely m a n y countable
405
models. The open mapping theorem for a stable theory states that for RAE = { p ( a ) E S(B) : p ( z ) does not fork over A } with A c B , the restriction map p E RAB ++ plA is an open mapping in the Stone topology. The following (OM) can be considered as a version of the open mapping theorem. Pillay’s proof in [14] uses (OM).
(OM) If tp(a/b) is semi-isolated and tp(a) is not isolated then a and b are dependent.
(OM) holds for stable theories and can be easily deduced from the open mapping theorem. Also, it can be proven directly, using the following characterization of independence: tp(a/b) does not fork over 0 iff tp(a/b) is finitely satisfiable in any model M 2 b. Unfortunately this characterization is not true for simple theories in the sense of Shelah [16]. The following example given by Kim illustrates the situation. Example 4.1. Let M be the { E ,R, C O , c1, ...}-structure defined by: (i) E is an equivalence class with two classes. Each class is infinite. (ii) For any two disjoint finite sets A and B contained in one class, there is a point d from the other class such that each point in A is R-related to d and no point in B is R-related to d. (iii) No two points in the same class are R-related. (iv) ti's are distinct elements contained in one class. Choose a 6 {ci : i < w } from the class containing the ti's. Choose b that is R-related to a but not R- related to any elements from {ci : i < w } . Then tp(a/b) is semi-isolated by R(a, b), but tp(a) is not isolated. Note that a and b have different types. However Kim succeeded to prove a supersimple version of Lachlan’s theorem without using (OM).
Fact 4.2. ( [ 5 ] ) Let T be a supersimple theory. T h e n T i s w-categorical, o r T has infinitely m a n y countable models. For proving his theorem, instead of using (OM), Kim showed the following fact for simple theories:
(*) Suppose that a and b are two independent realizations of a given type. Then tp(a/b) is semi-isolated if and only if tp(b/a) is semiisolated. However, examining his proof of the fact above, in fact we can show
(OM) if a and b have the same type (see Fact 4.3). The following ter-
406
minology is used in Fact 4.3. Let K be an infinite cardinal. We will say that a (partial) type is n-isolated over A if the type is implied by a consistent set of L(A)-formulas of cardinality < n. If p ( z ) is a complete type over A, p ( x ) will be called n-isolated if it is n-isolated over A. A complete type p ( x ) will be called (6, A0)-semi-isolated if there is a subtype q ( x ) with 1q(x)1 < 6 such that q ( x ) proves plA0. In this case we also say that q ( x ) (n,A)-semi-isolates p ( x ) . In the remainder of this section, T is simple.
Fact 4.3. ([20]) Let p ( x ) E S(A) be a non-n-isolated type. Let a and b be two realizations of p ( z ) such that tp(a/Ab) is (6, A)-semi-isolated. Then a and b are dependent over A.
Remark 4.1. Using the independence theorem for simple theories, we can show a little bit stronger result: Let p ( z ) E S(0) be a type such that p ( x 1 )U . . . Up(x,) has a finitely many completions. Let n 1 elements bl, ..., b,, a be independent realizations of p ( x ) . Suppose moreover that b l , ..., b, have the same Lascar strong type. If tp(a/bl, ..., b,) is semi-isolated then tp(a) is isolated
+
Proposition 4.1. Let [A]5 n. Let p ( x ) E S(A) be a non-6-isolated type. (1) Let I be a finite Morley sequence of p ( x ) . Then there is a model M 3 A in which I is a maximal Morley sequence over A. If n = No, we can find such a model even for a Morley sequence of infinite length. (2) Suppose that T is supersimple. Let a realize p ( x ) . Then there is a model M 3 Aa with dimN(p) = 1 (i.e. any two realizations of p ( x ) in M are dependent over A). (3) Let I = {ai : i E w} be a Morley sequence of p ( x ) . For each n E w there is a model M 3 A U {ai : i < n } that has no indiscernible sequence {bi : i < an} over A with tp({ai : i < n}/A) = tp({bi : i < n}/A). (4) 051) There is no supersimple theory T with 1 < I ( w , T ) < w. Proof. (1). By Theorem 2.2.19 in [2], it is sufficient to prove the following statement by induction on n 2 1: (*) If {ai : i 5 n} is a Morley sequence of p(x), then r(z) = {{ai : i 5 n} U { x } is a Morley sequence} is not 6-isolated. . Suppose that r ( x ) is 6-isolated. Let b realize r ( x ) . Then both b and a, realize the type q ( x ) = tp(a,/A, ao, ..., an-l). By the induction hypothe-
407
sis, q(z) is not K-isolated. On the other hand, tp(b/A, ao, ...,a,-l, a,) is ( K , (A, ao, ...,a,-1))-semi-isolated. So by Fact 4.3, b and a, are dependent over A, ao, ..., a,-1, contradicting the fact that {ai : i 5 n } U { b } is a Morley sequence. (2). Let A ( q ) be the following set of formulas:
p ( z )Up(y) U {z and y are independent over A}. Suppose that A(zy) is K-isolated over Aa. Then there are two realizations bo and bl such that (i) tp(bi/Aa) is ( K , A)-semi-isolated (i = 0 , l),and (ii) bo and b l are independent over A. From this, using a usual method, we can easily construct an independent (over A) sequence { b i : i = 1,2, ...} such that each tp(bi/Aa) is ( K , A)-semi-isolated. By Proposition 4.3, each bi and a are dependent over A. This contradicts the supersimplicity of T. So A(zy) is not K-isolated over Aa. Thus there is a model N 3 Aa which omits A(xy). (3). Clear by part (2). (4). We assume that T is small and non-w-categorical. Let p ( z ) be a nonprincipal type. Let {ai : i E w } be a Morley sequence of p . Let M , be a prime model over {ui : i < 2,). Then, by part (3), the M,’s are mutually 0 non-isomorphic.
Question 4.1. Is the assumption of supersimplicity necessary in part (2)? 5. An approach by Sudoplatov. Sudoplatov [17] used his geometric object called a polygonometry in a n approach to constructing a stable theory T with 1 < I ( w ,T ) < w . Roughly speaking, a polygonometry is a triple (G, (P,L , E), go) where G is a group, (P,L , E) is a pseudoplane and go is a fixed element in G. G is used to measure both the distance between two points in P and the angle between two lines in L. A trigonometry is a special kind of polygonometry such that any n-gon is expressed by a joint of 3-gons. He defines a certain language and considers a trigonometry (more precisely, the set P of points) as a structure for this language. He is trying to construct a stable theory T, with I ( w , T,) = n by applying Hrusovski’s generic construction method to his trigonometries. So far he has not obtained a complete counterexample, but his object seems to be a good’candidateof a tool for constructing such examples.
408
References 1. M. Benda, Remarks on countable models, Fund. math. 81 (1974), no. 2, 107-119. 2. C. C. Chang and H. J. Keisler, Model Theory 3rd ed., North-Holland, 1990. 3. E. Hrushovski, Finitely based theories, The Journal of Symbolic Logic, vo1.54 (1989), 221-225. 4. K . Ikeda, A. Pillay and A. Tsuboi, On theories having three countable models, Mah. Log. Quart. vol. 44 (1998), 161-166. 5. B. Kim, On the number of countable models of a countable supersimple theory. J. London Math. SOC.(2) 60 (1999), no. 3, 641-645 6. B. Kim, A note on Lascar strong types in simple theories, J. Symbolic Logic 63 (1998), no. 3, 926-936. 7. B. Kim and A. Pillay, Simple theories, Annals of Pure and Applied Logic, V O ~ .88 (1997), 149-164. 8. A. H. Lachlan, On the number of countable models of a countable superstable theory, Logic, methodology and philosophy of science IV (Proceedings, Bucharest 1971), North-Holland, Amsterdam (1973), 45-56. 9. D. Lascar, Ranks and definability in superstable theories, Israel Journal of Mathematics, vol. 23 no. 1(1976), 53-87. 10. T. Millar, Finite extensions and the number of countable models, The Journal of Symbolic Logic, vol. 54 (1989), 264-270. 11. M. G. Peretyat’kin, Theories with three countable models, Algebra i Logica, V O ~ .19 (1980), 224-235. 12. A. Pillay, Instability and theories with few models, Proc. Amer. Math. SOC. vol. 80 (1980), 461-468. 13. A. Pillay, Countable models of stable theories, Proceedings of American Mathematical Society 89 (1983) 666-672. 14. A. Pillay, An introduction to stability theory, Clarendon press, 1983 (Oxford logic guides:8). 15. S. Shelah, Classification Theory (2nd edition), North-Holland, 1990. 16. S. Shelah, Simple unstable theories, Annals of Mathematical Logic 19 (1989), pp. 177-203. 17. S. V. Sudoplatov, Group polygonometries and related algebraic systems (an informative survey), Contributions to General Algebra 11. Klagenfurt: Verlag Johannes Heyn (1999), pp. 191-210. 18. S. V. Sudoplatov, w-stable trigonometries on a projective plane, Siberian Advances in Mathematics, vol. 21 no. 4 (2002), pp. 1-28 19. A. Tsuboi, On theories having a finite number of nonisomorphic countable models, The Journal of Symbolic Logic, vol. 50 (1989), 806-808. 20. A. Tsuboi, Isolation and dependence in simple theories, in Proceedings of Model Theory at St. Catherine’s College Kobe Institute (1998). 21. R.L. Vaught, Denumerable models of complete theories, Infinitistic methods: proceedings of the symposium on foundations of mathematics, Warsaw, Pergamon Press (1961), 303-321.
409
PROBABILISTIC LOGIC PROGRAMMING WITH INHERITANCE*
JIE WANG
't3,
SHIER JU1 , XUDONG LUO 1,2 AND JUELIANG HU
Institute of Logic and Cognition, Zhongshan University, Guangzhou 510275, P. R. China Email: hspOOwjQstudent.asu.edzl.cn, [email protected] Department of Electrics and Computer Science, University of Southampton, Highfield, Southampton SO1 7 lBJ, United Kingdom Email: [email protected] Department of Philosophy, Beijing Normal University, Beijing, 100875, P.R. China School of Faculty of Science Zhejiang Institute of Science and Technology, Hangzhou 310033, P.R.China Email: hjlQzist.edu.cn
The paper proposes a new knowledge representation language that extends our probabilistic logic programming language [29], by using notation of inheritance. By introducing the mechanism of inheritance into the language, default reasoning with exceptions can be naturally represented but no inconsistent conclusions will be drawn. We illustrate these by encoding some non monotonic problems in the extended language. Its declarative model-theoretic semantics is developed, which is shown to generalise the answer sets semantics of probabilistic logic programs. We also analyse its complexity and show that inheritance does not cause any more computational overhead (i.e., its reasoning has exactly the same complexity as that in the probabilistic logic programming [29]).
'this work is supported by the key research project of Chinese ministry of education under grand no. 09050-3142003.
410
1. Introduction
As an important formalism for reasoning with classical knowledge, logic programming [14,1] started in the early 1970’s [12] based on earlier work in automated theorem proving, and began to flourish especially with the spreading of PROLOG. Now logic programming becomes a well-established knowledge representation and reasoning formalism in artificial intelligence and deductive databases. Further, since many applications in practice require knowledge representation and reasoning system to be able to deal with uncertainty, many efforts [21,22,23,24,5,15,16,17,18]have been made for integrating logic programming frameworks and probability-based representation and reasoning formalisms. Probabilistic logic programs are now widely recognized as a valuable tool for knowledge representation and commonsense reasoning of probabilistic information. In [29], we also presented a probabilistic logic programming language. In particular, we extend the answer set semantics of Gelfond and Lifschitz [7] by associating a probability with each atom in a program in this language. However, sometimes probabilistic inconsistencies can generate in this framework. For instance, consider the famous nonmonotonic reasoning example: 80% of all birds can fly while penguins do not fly. Assume that our knowledge about a particular individual, called tweety, is that “tweety is a bird” and “tweety is a penguin”, we conclude “tweety can fly with the possibility 80%” and “tweety can not fly” as well. This is inconsistent. In order to remove this kind of inconsistence, this paper integrates the technique of reference-class reasoning [28,13,26] into our probabilistic logic programming language [29]. That is, to take the following ideas behind reference-class reasoning into our language: any class containing the particular individual can be considered as reference class, and the smaller reference classes are preferred to larger ones. More precisely, this paper proposes an extension of our probabilistic logic programming language [29] by adding some machinery for inheritance with overriding from default reasoning, denoted PLP<. That is, a subclass inherits all properties from its superclass, but a subclass may refine some of the inherited properties and introduce extra definitional properties local to itself. Possible probabilistic inconsistencies are resolved in favor of the rules which are “more specific” according to the inheritance hierarchy. In this way, a direct and natural representation of default reasoning with exceptions is achieved. It turns out that such an approach can be developed as a probabilistic generalization of classical default reasoning.
41 1
The rest of the paper is organized as follows. In Sections 2 and 3, we formally define our PLP< language based on probabilistic logic programming [29], and provide a declarative model theoretic semantics of PLP<. Section 4 illustrates the knowledge modelling features of the language by encoding classical nonmonotonic problems in PLP<. In Section 5 , we analyse the computational complexity of reasoning over PLP< programs. Section 6 discusses the relate work. Finally, Section 7 summarises the paper and pout out the direction of further research. 2. Syntax of PLP<
Let q5 be a first order vocabulary that contains a finite set of predicate symbols and finite set of variable symbols. Let Q be a set of object constants and ,B be a set of probabilistic constants. Object constants represent elements of a certain domain, while probabilistic constants represent real number in the unit interval [0,1]. A term is either a variable or an object constant. An atom is a construct of the form A(t1,.. . ,tn),where A is a predicate of arity n in q5 and t l , . . . , tn are terms. In addition, in PLP< language, we define a finite partially ordered set of symbols (A, <), where A is a set of strings, called object identifiers, and < is a strict partial order (i.e., the relation < is: (i) irreflexive, c { c, VCE A; and (ii) transitive, (u < b) A ( b < c ) (a < c ) , Vu, b, c E A). In order to manipulate probability information, we introduce a new notation:
*
Definition 2.1. A special literal patom is an expression of the form L / p , where L is an atom, and p E [0,1] is the probability that L occurs. Intuitively, when p = 1, L/1 means L is true (we abbreviate L / 1 by L ) ; when p = 0, L/O means 1 L (i.e., L is false); and when p E [0,1], L / p means L probably occur (i.e., L is probabilistic true). Thus, if toss is a unary predicate in our language, toss(head)/0.5 means that the probability that a tossing a coin yields head is 0.5. Definition 2.2. A rule
T
is an expression of the form:
LO/PO + L I / P I ,.. . , L m / ~ mnot , Lm+l/~rn+l,.* . ,not Ln/Pn,
(1)
where n 2 rn 2 0; each Li is an atom; 0 _< pi _< 1, and Li/pi (0 5 i _< n) is a patom; not is a connective called negation as failure or default negation, and not L / p means there is no reason to believe that the probability that L occur is p E [0,1]. The expression of the form LO/po is the head of
412
r. L l l p l , . . . ,L,/p, is called the positive part of the body of r and not L,+llp,+l,. . .,not L J p , is called the negation as failure part of the body of r. We often denote the sets of patom appearing in the head, in the positive part of the body, and in the negation as failure part of the body of a rule r by Head(r),Body+(r)and Body-(r), respectively. So we abbreviate rule (1) by Head(r) t Body+(r),Body-(r).
(2)
The intuition of the above rule is: if the probability that Li is believed true is pi ( 1 i m) respectively, and have no reason to believe in L,+I,. . . ,L , with the probability p m + l , . . ., p n respectively, then the probability Lo is believed true is PO. In the context of probabilistic logic programming with inheritance, there are two major kinds of negation: the first one is classical negation, given that the probabilistic of an atom L is p, it is easy to verify that the probability of E , the classical negation of L is 1- p, the second kind of negation is default negation not L.
< <
Definition 2.3. An object o is a pair < oid(o),C(o) >, where oid(o) is an object identifier in A and C ( o ) is a (possible empty) set of rules. A knowledge base on A is a set of objects, one for each element of A. The relation < induces a partial order on program II in the obvious way, that is given oi = (oid(oi),C ( o i ) ) and oj = (oid(oj),C ( o j ) ) ,oi < o j , iff d d ( o i ) < oid(oj) (read “oi is more specific than o j ” ) . Informally, a knowledge base can be viewed as a set of objects embedding the definition of their properties specified through probabilistic logic rules.
Definition 2.4. Given a knowledge base k and an object identifier o E A, the PLP< program for o on k is the set of objects II = {(o‘,C(o’)) E k I o = 0‘ or o < 0’). Thank to the inheritance mechanism, a PLP< program II for an object o incorporates the knowledge explicitly defined for o plus the knowledge inherited from the higher objects. If a knowledge base admits a bottom element (i.e., an object less than all the other objects, by the relation <), we usually refer to the knowledge base as “program”. Since it is equal to the program for the bottom element. Moreover, we represent the transitivereduction a of the relation < on the object. In order to represent the a(a,b) is in the transitive-reduction of
< iff a < b and there is no c such that a < c and
413
numbership of a pair of objects (resp., object identifiers) ( 0 2 , o l ) to the transitive-reduction of < we use the notation 0 2 : 01 , and say that 0 2 is a sub-object of 01.
Definition 2.5. A term, an atom, a p-atom, a rule, or a PLP< program is ground if and only if it does not contain any variable symbols. Example 2.1. Consider the following program 11: (PhD-student)
{get-PhD(Jane)/0.8 t PhD-student( Jane), not master-student (Jane), PhD-student (Jane)10.9 t} oz(Jane) : ol(PhD-student) {get-PhD(Jane)/O t, has-publication( Jane) t PhD-student( Jane)}
01
Clearly, II consists of two object 02 and 01 , 0 2 is a sub-object of 01. Since Jane is just probably a PhD student, it is possible that she is a bachelor student. If she is a a bachelor student indeed, it is impossible for her to get a PhD. So, the above program is reasonable. 3. Semantics of PLP<
Suppose that a knowledge base k is given and an object o has been fixed. Let II be a PLP< program with respect to o on k. The universe an of II is the set of all object constants and ,On is the set of all probabilistic constants appearing in the rules. The basic base bn of II is the set of all possible ground atoms constructible from the predicates appearing in the rules of II and the object constants occurring in an. For an arbitrary set S , its power set exists and is the set of elements T ( S )= {z 1 z C S } .
Definition 3.1. For an arbitrary set S and its power set T(S). If A E T ( S ) , product of the set A is the product of all elements which belong to A, denoted IAl. Product set of S is
4 s )= (1-41 I A E 491.
c < b.
(3)
414
The em base bn of II is the set of all possible ground p-atom constructible from the atoms appearing in the basic base bn and the probabilistic constants occurring in J ( ~ u ) Note . ~ that, the Base of a PLP< program contains all possible combination of the basic base bn and J ( p n ) . Given a rule r occurring in II, a ground instance of r is a rule obtained from r by replacing every variable X in r by o ( r ) , where o is a mapping from the variables occurring in r to the object constants in the rules occurring in an. We denote by grovnd(II) the finite multiset of all instances of the rules occurring in II. The reason why grovnd(II) is a multiset is that a rule may appear in several different objects of II, and we require the respective ground instances be distinct. Hence, we can define a function obj- of from ground instance of rules in g r m n d ( I I ) onto the set A of the object identifiers, associating with a ground instance F of r (i.e., the unique object of T).
Definition 3.2. A subset of ground patom in II is said to be consistent if whenever p-atom of the form L / a and L i b are in it simultaneity, then a = b. An interpretation I is a consistent subset of Bn. Definition 3.3. Given an interpretation I E Bn, if L / p E I holds (denoted I k L i p ) , a ground patom L i p is true w.r.t. I; otherwise, L i p is false w.r.t. I. Definition 3.4. Given a rule r E g r m n d ( I I ) , the head of r is true in I if Head(r) E I . The body of r is (probabilistic) true in I if (i) every patom in Body+(r) is true w.r.t. I , or there exist patom L / 1 in Body+(r) and L / l E I , but 3p E ( O , l ) , L / p E I ( 1 5 i 5 m ) ; and the other p-atom in Body+(r)is true w.r.t. I ; and (ii) every p-atom in Body-(r) is false w.r.t. I. Definition 3.5. Given a rule r E grovnd(II),rule r is satisfied in I if either the head of r is true in I or the body of r is not true in I . Definition 3.6. Given a rule r E ground(II), Head(r) = Lo/po, rule r is probabilistic satisfied if (i) the body of r is (probabilistic) true in I , without loss of generalisation, assume L i / l , . . . ,L j / l E Body+(r) and Lilai,. . . , L j / a j E I (ah # 1 , k = i,. . . , j ) ; (ii) every atom in Body+(r)is probabilistic independent of each other; and (iii) Lo/(po x ai x . . . x a j ) E I. bThis is because the element of /3n belong to [0, 11, so all element of J(j3n) also belong to [O, 11.
415
Definition 3.7. Two ground rules r1 and 7-2 are inconsistent on L if L / a is the head of rule T I and L/b is the head of rule r2 but a # b. In particular, two ground rule r1 and 1-2 are conflict on L if L / 1 is the head of rule r1 and L/O is the head of rule r2. Next we introduce the concept of a model for a PLP< program. Unlike in traditional logic programming, the notion of satisfactory of rules is not sufficient for this goal because it does not take into account the present of explicit inconsistency. Hence, we first present some preliminary definitions.
Definition 3.8. In an interpretation I , for two arbitrary ground rules q , r 2 E ground(II), suppose that L / a is the head of rule 7-2. Then we say that T I overrides 7-2 on L in I if (i) obj-of(r1) < o b j - o f ( r ~ )(ii) , 3b E [0, 11, L / b E I ( a # b), and (iii) the body of r2 is (probabilistic) true in I . A rule r E ground@) is overridden in I , if for the head L / a of r there exists r1 in ground(II) such that r1 overrides r on L in I . Intuitively, the notion of overriding allows us to solve inconsistent or conflict arising between rules with inconsistent even consistent head. For instance, suppose that both L / a and L / b ( a # b ) are derivable in I from rules r and r’, respectively. If T is more specific than T ‘ in the inheritance hierarchy, then r’ is overruled meaning that L / a should be preferred to L / a because it is derivable from a more trustable rule.
Example 3.2. Consider the program IT of Example 1. Let I = {get-PhD(Jane)/O,has-pubZication(Jane)/O.S, PhD-student(Jane)/0.9} be an interpretation. Rule
get-PhD(Jane)/O c in the object
02
overrides rule
get-PhD( Jane)/0.8 t PhD-student( Jane), not master-student(Jane) in
01
on the atom get-PhD(Jane) in I .
Example 3.3. Consider the program 01
(PhD-student)
oz(Jane) : 01
IT0
{get-PhD( Jane)/0.8 t PhD-student( Jane), not master-student( Jane), PhD-student(Jane)/0.9 t} (PhD-student) {get-PhD(Jane)/0.8t, has-publication( Jane) t PhD-student( Jane)
416
Let I = {get-PhD(Jane)/0.8, has-publication( Jane)/O.S, PhD-student(Jane)/0.9}, the first rule, rl, of 01: and the first rule, 7-2, of o2 have the same head get-PhD(Jane)/0.8, because the bodg+(rl) of is (probabilistic) true in I . So inconsistency generates. Rule 7-2 overrides rule 7-1 on the atom get-PhD(Jane) in I . Definition 3.9. Let I be an interpretation for II. I is a model for II if each rule in ground(Il) is satisfied, probabilistic satisfied or overridden in I . I is a minimal model for II if no proper subset of I is a model for II.
Definition 3.10. Given an interpretation I for IT, the reduction of II w .r .t. I , denoted by Gr(II), is the set of rules obtained from ground(II) by: (i) removing each rule that is overridden in I , (ii) removing each rule T that has property Body-(r) n I = 4, and (iii) removing the negation as failure part from the bodies of the remaining rules. Example 3.4. Consider the program 11 of Example 1. Let its interpretation I = {get-PhD(Jane)/O, has-publication( Jane)/O.S, PhD-student( Jane)/O.9}. (4) As shown in Example 3, the first rule in 01 is overridden in I . Thus, Gl(II) is the set of rules {get-PhD(Jane)/O c , PhD-student(Jane)/0.9 t, has-publication( Jane) t PhD-student( J a n e ) } . Consider now another interpretation
M = {get-PhD(Jane)/O,PhD-student(Jane)/O.S}. It is easy to see that Gn.r(II) = Gr(II).
(5)
We can see that the reduction of a program is simply a set of ground rules that do not contain not. Definition 3.11. Let M be a model for II. We say that M is an (PLP<) answer sets for II if M is a minimal model of the Gn.r(II).
Example 3.5. Consider the program II of Example 1. It is easy to see that the interpretation (4) is a minimal model for GI@).Moreover, it can be easily realized that I is the only answer set for II. On the other hand,
417
the interpretation ( 5 ) is not an answer set because it is not a model for GM(II) since the the second rule in 0 2 in ground(GM(II)) is not satisfied or probabilistic satisfied.
4. Knowledge representation with PLP< In this section, we present a number of examples which illustrate how knowledge can be represented using PLP<. To start, we show that PLP< encodes the famous nonmonotonic reasoning example stating that 80% of all birds can fly while Penguins do not fly.
Example 4.1. Consider the following program I'I with A(II) consisting of three objects bird, penguin and tweety, such that penguin is a sub-object of bird: bird ( f l y ( z ) / 0 . 8t bird(z)} penguin : bird { f l y ( z ) / O t}. Unlike in probabilistic logic programming [8], we may derived a pair of inconsistent ground patom fZy(penguin)/0.8 and f l y (penguin)/O from the set of rules. According to the inheritance principles, the ordering relationship between the objects can help us to assign different levels of reliability to the rules, and thus allow us to solve possible inconsistency or conflicts. For instance, in our example, the inconsistent conclusion penguin both possibly flies and does not fly to be entailed from the program (as penguin is a bird). However, this is not the case. Indeed, the lower rule fZiesCpenguin)/O t specified in the object penguin is considered as a sort of refinement to the general rule, and thus the meaning of the program is rather clear: tweety probably flies and penguin does not fly. That is, flies(penguin)/O t is preferred to the rule fZies(penguin)/0.8 t as the hierarchy explicitly states the specificity of the former. Intuitively, there is no doubt that M = {flies(tweety)/0.8,fZies(penguin)/O} is the only reasonable conclusion. Let us see another example.
Example 4.2. Consider the following knowledge base representing a set of security specification about a simple part-of hierarchy of objects. {TI
01
: authorize(Bob)/0.9 t not
authorize(Ann)/l, authorize(Ann)/0.6 t not authorize(AZice)/O} : authorize(AZice)/O t} : authorize(Bob)/O t}
r2 : 02
03
: 01 : 01
(7-3 (7-4
418
Objects 0 2 and 03 are, respectively, a part of the object 01. Access authorisations to objects are specified by rules with head predicate authorise and subject to which authorisations are granted appear as arguments. Inheritance implements the automatic propagation of authorisations from an object to all its sub-objects. The overriding mechanism allows us to represent exceptions: for instance, if an object o inherits a possible authorisation with probability a but another probability b for the same subject is specified in 0 , then the latter authorisation prevails on the former one. Consider the program 11 03 = ( ( 0 1 , (7-1,T Z } ) , (02, (7-3))) for the object 0 2 on the above knowledge base. This program defines the access control for the object 0 2 . Thanks to the inheritance mechanism, authorisations specified for the object 01, to which 0 2 belongs, are propagated also to 0 2 . It consists of rules 7-1, 7-2 (inherited from 01) and 7-3. Ruler 7-1 states that Bob is authorised with probability 0.9 to access object 0 2 provided that no authorisation for Ann to access 0 2 exists. Rule 7-2 authorises Ann with probability 0.6 to access 0 2 provided the fact that Alice is authorised with probability 0 to access 0 2 is not derive. Finally, rule 7-3 defines a denial for Alice to access object 0 2 . Due to the absence of authorisation for Ann, the authorisation to Bob of accessing the objects 0 2 with probability 0.9 is derived by rule T I . Further, the explicit denial to access the object 02 for Alice (rule 7-3) does not allow us derive the authorisation for Ann with probability 0.6 by rule 7-2. Hence, the only answer set of this program is {authorize(Bob)/0.9, authorize(Alice)/O}. Consider now the program II 03 = { ( 0 1 , { 7 - 1 , ~ } ) , ( 0 3 , { 7 - 4 } ) } for the object 0 3 . Rule 7-4 defines a denial for Bob to access object 0 3 . The authorisation for Bob with probability 0.9 defined by rule 7-1 is no longer derived. Indeed, even if rule 7-1 allow us derive such an authorisation due to the absence of authorisation for Ann, it is overridden by the explicit denial (rule 7-4 ) define in the object 0 3 , i.e., at a more special level. The body of rule 7-2 inherited from 01 is true for the program. Since no denial for Alice can be derived. The program IT 0 3 has answer set {authorise(Bob)/O, authorise(Ann)/0.6}.
5. Computation complexity In this section, firstly we analyse the computation complexity of the decision problems corresponding to Brave Reasoning in the propositional case: Given a PLP< program II and a patom L / p , decide whether there exists an answer set M for IT such that L/p is true w.r.t. M .
419
We next prove that the complexity of reasoning in PLP< is exactly the same as in probabilistic logic program [29]. That is, inheritance comes for free, as the inheritance adding does not cause any computational overhead. For the sake of space, we discuss this only in the propositional case (i-e., the case of ground PLP< programs).
Lemma 5.1. Given a ground PLP< program II and an interpretation M for II, deciding whether M is an answer set for II is in CoNP. Proof. We check in NP that M is not an answer set of II as follows. Guess a subset I of M , and verify that: (i) M is not a model for GM(II),or (ii) I is a model for GM(II) and I c M . The construction of GM(II) is feasible in polynomial time, and the tasks (i) and (ii) are clearly tractable. Thus, deciding whether M is not an answer set for II is in NP, and consequently, deciding whether M is an answer set for II is in CoNP. In order to prove the following theorem, we need two more lemmas. In [7], Gelfond and Lifschitz propose answer set semantics which is the most widely acknowledged semantics. We also defined the semantics of probabilistic logic programming [29] which is a variant of answer set semantics, for this reason, while defining the semantics of our language, we took care of ensuring full agreement with answer set semantics on inheritance-free programs.
Lemma 5.2. Let 11 be a PLP< program consisting of a single object o = (oid(o),C(o)). Then M is an answer set of ll if and only if it is a consistent answer set of C(o) (according to [29]). Lemma 5.3. Probabilistic logic programming under answer sets semantics is ng-complete. Proof. The semantics of probabilistic logic programming defined in [29] is a variant of answer set semantics proposed by Gelfond and Lifschitz [8]. The main difference is lie in the conditional transformation in step 1 in Section 3 which is in NP, while extended logic programming under answer sets semantics is Ci-complete [4]. So the problem is IIg-complete. Theorem 5.1. Brave reasoning on PLP< programs is Cg-complete. Proof. Given a ground PLP< program II and a ground p a t o m L l p , we verify that L / p is a brave consequence of 11 as follows. Guess a set M 5 Bn of ground patom, check that (i) M is an answer set for II, and (ii) Llp is true w.r.t. M, Task (ii) is clear polynomial; while (i) is in CoNP, by virtue
420
of Lemma 1. The problem therefore lies in C;. Further, Cg-hardness follows 0 from Lemmas 2 and 3. So, the theorem holds.
6. Related work
The work that is the closest to this paper is the one by Buccafurri et al. [3]. It describes a knowledge representation language to disjunctive logic programs with inheritance. The disjunctive logic programming only deals with disjunctive knowledge and non monotonic negation, while probabilistic logic programming aims handle with numerical uncertainty. In other words, [3] does not handle uncertainty, but ours does. Lukasiewicz [17] expresses an intention similar to ours, where he also discussed probabilistic logic programming under inheritance with overriding (PLPIO). In fact, he investigated a kind of approach of strengthening the notion of logic entailment in probabilistic logic programming. The approach is inspired by reference-class reasoning to Reichenbach [28] that can be interpreted as inheritance with overriding as it is known from objectoriented programming languages. There the notation of inheritance means that any class containing the particular individual can be considered as reference class, while the notation of overriding means that smaller reference classes are preferred to larger ones. However, the work is different from ours in a number of aspect. First, the language of PLPIO does not include default negation 1,while our PLP< includes it. Second, PLP< and PLPIO present different framework of probabilistic logic programming, and so they have completely different semantics. Third, our PLP< generalises consistent answer set semantics to probabilistic logic programming with inheritance, while PLPIO does not.
7. Conclusion This paper presents the syntax and semantics of a probabilistic logic programming language with inheritance, which is based on recent approach to probabilistic logic programming [29]. Moreover, we analyse the computational complexity of probabilistic logic programming with inheritance, and conclude that inheritance does not cause any more computational complexity (i.e., reasoning in PLP< has exactly the same complexity as reasoning in probabilistic logic programming). In the future, we plan to investigate the applications of this framework beyond default reasoning. It must be interesting to apply our work (that
42 1
combines logic and uncertainty) in the active research area of agent-based automated negotiation. This is possible since there are already some work (e.g., [27,10,11]) that bridges logic and agent-based automated negotiation, a n d some work t h a t bridges uncertainty and agent-based automated negotiation [19,20]. References 1. K. R. Apt. Logic programming, Handbook of Theoretical Computer Science, volume B, Chapter 10,pp. 493-574,MIT Press, 1990. 2. F. Bacchus, A. Grove, J.Y. Halpern, and D. Koller. From statistical knowledge bases to degrees of belief, Artificial intelligence, 87, pp. 75-143,1996. 3. F. Buccafurri, w. Faber, N. Leone. Disjunctive logic programs with inheritance. Theory and practice of logic programming, 2 (3): 293-321,2002. 4. E. Dantsin, T. Eiter, G. Gottlob and A. Voronkov. Complexity and expressive power of logic programming. In proceedings of the Twelfth annual IEEE conference on computational complexity, June 24-27,pages 82-101,1997, 5. M. I. Dekhtgar, A.Dekhtgar and V. S. Subrahmanian. Hybird probabilistic programs: Algorithms and complexity. In proceedings UAI-99, pages 160-
169,1999. 6. R. Fagin, J. Halpern and N. Meggido. A logic for reasoning about probabilities. Information and Computation, 87: 78-128,1990. 7. M. Gelfond and V. Lifschitz. The stable model semantics for logic programming. Logic programming: Proc. of the Fifth Int. Conf. and Symp., pages
1070-1080,1988. 8. M. Gelfond and V. Lifschitz. Classical negation in logic program and disjunc-
tive databases. New generation computing, pages 365-387,1991. 9. J. Y.Halpern. An analysis of fist-order logics of probability. Artificial Intelligence, 46(3): 311-350,1990. 10. M. He and N.R. Jennings. SouthamptonTAC: Designing a Successful Trading Agent. Proceedings of the Fifth European Conference of Artificial Intelligence, pages 8-11,2002. 11. M.He, H.F. Leung, and N.R. Jennings. A fuzzy logic based bidding strategy in continuous double auctions. IEEE Transactions on Knowledge and Data Engineering. 2003. To appear. 12. R.A. Kowalski. Predicated logic as a programming language. In information processing’74, pages 569-574,1974. 13. H. E. Kybur. Jr.The reference class. Philos.Sci,50:374-397,1983. 14. J. Lloyd. Foundations of logic programming. Springer, 1984. 15. T. Lukasiewicz. Probabilistic logic programming. In proc of the 13th Beinnial European Conf. On Artificial Intelligence, pages 388-392,1998. 16. T.Lukasiewicz. Many-Valued Disjunctive Logic Programs with Probabilistic Semantics. Proceedings of the 5th International Conference on Logic Programming and Nonmonotonic Reasoning, Volume 1730 of LNAI, pages 277-289,
1999.
422
17. T. Lukasiewicz. Probabilistic logic programming under inheritance with overriding. Technical report IN-FSYS RR-1843-01-05, Institut fur Information SSYS-teme, TU wien, 2001. 18. T. Lukasiewicz. Probabilistic logic programming with conditional constraints. ACM trans. Computat, logic 2(3), pages 289-337, 2001. 19. X. Luo, C. Zhang, and N.R. Jennings. A Hybrid Model for Sharing Information between Fuzzy, Uncertain and Default Reasoning Models in Multi-Agent Systems. International Journal of Uncertainty, Fuzziness and KnowledgeBased Systems 10(4), 401-450, 2002. 20. X. Luo, N.R. Jennings, N. Shadbolt, H.F. Leung, and J.H.M. Lee., A fuzzy constraint based model for bilateral, multi-issue negotiation in semicompetitive environments. Artificial Intelligence, 2003. To appear. 21. R. Ng and V. S. Subrahmanian. Probabilistic logic programming. Information and computation, lOl(2): pages 150-201, 1992. 22. R. Ng and V. S. Subrahmanian. A semantical framework for supporting subjective and conditional probabilities in deductive databases. J. autom. reasoning, lO(2): pages, 191-235, 1993. 23. R. Ng and V. S. Subrahmanian. Stable Semantics for Probabilistic Deductive Databases, Information and Computation, 110 (l), pages 42-83. 1995. 24. L. Ngo, P. Haddawy. Probabilistic Logic Programming and Bayesian Networks. ASIAN 1995, pages 286-300, 1995. 25. N. J. Nilsson. Probabilistic logic. Artificial Intelligence, 28(1): 71-87, 1986. 26. J. L. Pollock. Nomic probabilities and the foundations of induction. Oxford university press,1990. 27. S. Parsons, C. Sierra, and N.R. Jennings. Agents that reason and negotiate by arguing. Journal of Logic and Computation. 8(3), 261-292, 1998. 28. H. Reichenbach. Theory of probability. University of California Press, Berkeley, CA, 1949. 29. J. Wang and S. Ju. Probabilistic logic programming. Proceedings of the 8th joint international computer conference. Zhejiang University Press, 2002.
423
SEQUENT SYSTEMS FOR CLASSICAL AND INTUITIONISTIC SUBSTRUCTURAL MODAL LOGICS
OSAMU WATARI Division of Systems and Information Engineering, Graduate School of Engineering, Hokkaido University, Kita 13 Nishi 8, Kita-ku, Sapporo 060-8628, Japan E-mail: [email protected]
TAKESHI UENO Department of Mathematics, Hokkaido University, Kita 10 Nishi 8, Kita-ku, Sapporo 060-0810, Japan
KOJI NAKATOGAWA Department of Philosophy, Hokkaido University, Kita 10 Nishi 7, Kita-ku, Sapporo 060-0810, Japan
MAYUKA F. KAWAGUCHI AND MASAAKI MIYAKOSHI Division of Systems and Information Engineering, Graduate School of Engineering, Hokkaido University, Kita 13 Nishi 8, Kita-ku, Sapporo 060-8628, Japan
We introduce sequent style inference rules for classical and intuitionistic substructural modal logics. Using these inference rules, we define modal full Lambek calculi. We extend the algebraic interpretations, which have been given to the substructural (non-modal) logics in the previous investigations, so as to provide algebraic interpretations for modal full Lambek calculi. Using these interpretations, we proved soundness and completeness theorems for each of the modal full Lambek calculi.
1. Introduction
Among sequent systems for substructural logics, FL (full Lambek calculus) provides the basic framework for the investigation of substructural logics (see [4],151). Extending FL by adding structural inference rules (exchange, contraction, weakening), one obtains a series of substructural logics.
424
Addition of modal operator and their inference rules to the sequent system LK for classical logic led some researchers (Ohnishi, Matsumoto [3] and others) to the introduction of the sequent systems for classical modal logics. Intuitionistic versions of these systems are obtained by restricting the right hand side of a sequent in the well-known manner. Our main purpose here is to construct both classical and intuitionistic substructural modal sequent systems which result from FL and CFL, (classical full Lambek calculus with exchange) through the addition of modal operators 0 and 0 together with appropriate inference rules. (When we delete the exchange rule from classical systems, essentially different systems axe yielded through the change of the order of formulas in the left and right sides of sequents occurring in some inference rules [5]. Detailed investigations of these systems will be carried out in a separate paper. We consider here only extensions of CFL,.) We will prove soundness and completeness theorems for them through algebraic interpretations. 2. Formal system 2.1. Language, formula In this section, we introduce the language C for substructural modal logic. The basic propositional connectives of C consist of multiplicative conjunction * (called “asterisk”), multiplicative disjunction (called “plus”), additive conjunction A (called “and”), additive disjunction V (called “or”), implication 3 and >’, negation 1 and 4 , necessity 0 (called “box”), and possibility 0 (called “diamond”). We use t to denote the unit element for multiplicative conjunction. f denotes the unit element for multiplicative disjunction. T is called “top”, and it denotes the unit element for additive conjunction. I is called “bottom”, and it denotes the unit element for additive disjunction. As in FL, 1 A is an abbreviation of A > f , and 4 A is an abbreviation of A 3‘ f . For the propositional variables of C,we use symbols p o , p l , . . . . The set F of formulas of C is the smallest set which contains all the propositional variables of C as well as all the unit elements of C,and which is closed under the following rules:
+
(1) A E F (2) A , B E
* i A , - ‘ A , o A ,O A E F F
+ A * B , A + B , A A B , A v B , A 3 B , A 3‘ B E F
The classical substructural modal logics which we investigate in this
425
paper have the exchange rule. Under the presence of the exchange rule, there is no need to differentiate 3 and 3’.In classical cases, we therefore delete 3’.Since 4 A is defined as A 3’f , 4 is to be deleted from classical cases. It should be noted that the language of intuitionistic substructural modal logics do not contain the multiplicative disjunction In a sequent system, * corresponds to the “comma” on the left side of the sequent, and + corresponds to the “comma” on the right hand side. + is the dual of *. In intuitionistic sequent calculus, the “comma” never occurs on the right hand of the sequent. This is because, in an intuitionistic sequent system, one cannot write the inference rules for in such a way that the duality of and * is represented in inference rules. A sequent of substructural modal logics is a list of formulas having the following form:
+.
+
+
AI,.. .,A,
+ B1,. . . , B ,
where each Ai and Bj are any formula. Note that both sides of + is allowed to be empty. For the intuitionistic cases, no formula or just a single formula occurs on the right-hand-side of the sequent. 2.2. Inference rules
2.2.1. Classical substructural modal logics We introduce the system called CFL:, which is the basis for the other classical substructural modal logics. CFLF is obtained from the standard classical substructural logic CFL, [4,5] by adding the following inference rule (0-K):
or + AO A +
(0-K)
This rule was introduced and shown, in [3], to correspond to the axiom K for modality:
K : o ( A 3 B ) 1 ( O A 3 OB) Definition 2.1. (Inference rules for classical modal substructural logics) CFL: have the following axioms and inference rules: 0
Axioms and rules for logical constants:
A+A
426
-+t
r,rl-+ A -+ A tw
I',t,"I
f-+
r -+ A , A I + A , f , A l fw
Structural inference rule:
r , A ,B,r' -+ A (e left) r , B , A , F -+ A
r + A, A, B , A' r -+ A, B , A , A/
r -+ A , A r ' , A + A' r,rl-+A, A! 0
(e right)
(cut)
Logical inference rules:
7 -+ ~ A,-+,rAA r,A , ri -+ n r,AAB,P+A r,B , r1-+ A F , A A B , P -+ A
(1
left)
(A left) (A left)
r , A -+ A iA,A
r -+
right)
(1
r
-+ A,A,A' -+ A , B , A ' (A right) r -+ A , A A B,Al
427
The set of sequents provable in CFLq is defined to be the intersection of sets X’s each of which satisfies the following two conditions: 0 0
All the axioms of CFLE are members of X . If the upper sequent(s) of an inference rules of CFLE is in X , then its lower sequent is also in X .
The other systems, CFLfD, CFLqT, CFLfT4 and CFLFT5are obtained from CFLF by adding the following inference rules:
Or
A (0-4)
or -+ O A -+
These rules correspond to the following axioms for modality:
D : OA>iOiA T : OA>A 4 : OA>OOA 5 : i0-A > OTCI~A Among the above four rules, the first rule (0-D) is due to Valentini [7]. The rest of them are adopted from [3]. The provability for the above systems is defined similarly for the provability of CFLq
428
2.2.2. Intuitionistic substructural modal logics We introduce the basic system of intuitionistic substructural modal logics, called FLK. In case of classical modal logics, 0 and 0 are mutually definable with the help of 1.However, this mutual definability of 0 and 0 does not hold for intuitionistic modal logics[6]. As a result, FLK has the inference rule for the 0-modality and the inference rule for 0-modality. The classical modal logic K has two axioms: O ( A 3 B ) 3 ( O A 3 U B ) and O ( A 3 B ) 3 (OA 3 OB). The first axiom corresponds to the inference rule:
r+A
or + O A The second axiom corresponds to the inference rule:
A+C OA + OC These inference rules are investigated in [3]. In case of intuitionistic logic, the right-hand-side of the sequent is restricted to one or zero occurrence of a formula. Therefore, one cannot use the second inference rule for the 0-modality. Instead, we introduce two inference rules (O-Kl) and (O-Kz) for the +-modality:
as well as the (0-K) rule which is the same rule as in the classical cases. (The lack of the exchange rule forces us to device the inference rule for the 0-modality into two cases.) These three inference rules are used to introduce FLK.
Definition 2.2. (Inference rules for intuitionistic substructural modal logics) FLK have the following axioms and inference rules: 0
Axioms and rules for logical constants:
A+A
429
r,r/+ c t w r + fw r,t,r/-+c r+f 0
Structural inference rule:
rl + A r 2 , ~ ,+r 3c r2,r1,r3+ c 0
(cut)
Logical inference rules:
r + A B,r'+C r,A +B (3 left) (3 right) A 3 B,F,l-" + C r+A>B
r + B~, r / + c (3'left) r , A 3' B,r' + C A
left)
l ~ , -+ r +
r
(1
y A+ ~
+
left)
(1)
A,r + B (3' right) I'+AYB
r,A+
r+
A,F
right)
(1
l~
+
r + 1 ' (-2 ~
right)
r,A,r/+ c
(A left) F , A A B,I" + C r,B,r' + c (A left) r,AA B,P +C
(A right)
430
In the above axioms and inference rules, the formula C may not occur on the right-hand-side of a sequent. The other systems, FLKD,FLKTand FLKT4are obtained from FLK by adding the following inference rules:
These rules correspond to the following axioms for modality:
D : oA>OA T : UA>A,A>OA 4 : UA > DOA, OOA
> OA
The first rule (00-D) is due to Valentini[7]. The next three rules, namely (0-T),(0-T),and (0-4), are adopted from [3]. The last two rules, (0-41) and (0-42), are introduced here in order to circumvent the same sort of difficulty which we have faced with in the introduction of (O-K1) and (O-Kz). The provability for the above systems is defined similarly for the provability of CFLF 3. Algebraic interpretation
In this section, we will introduce algebraic structures, called modal fuZZ Lambek algebra. An algebra corresponding to (non-modal) full Lambek calculus was introduced in [2]. He called it “uni-residuated lattice-ordered groupoid” ,
431
which does not have an operation corresponding to 3’. Ono[4] introduced an “FL-algebra” where both 3 and 3’ are took into consideration. Furthermore, Ono extended it by adding the modalities (!, ?) for linear logics. They were called “modal full Lambek algebras” [4]. We will investigate modal full Lambek algebras for modalities such as K, K D , KT, KT4 and KT5. Definition 3.1. (Modal full Lambek algebras) A structure A = (V,U, n, 0,+,+’,L, M , 1,0,T, I)is a FLK-algebra if the following conditions are satisfied: (1) (V, U, n,0,+ , + I ,
1,0,T, I)is a FL-algebra:
(a) (V,U, n, T, I)is a lattice with the least element I and the greatest element T for which T = I -+ I holds, (b) (V,0 ,1) is a monoid with the identity 1, (c) Vx,y,z,wEV z o ( x u y ) o w = ( z o x o w ) u ( z o y o w ) , (d) Vx,y, z E V ((x o y 5 z e x 5 y + z ) and (x o y 5 z u y 5 IC +‘z ) ) , (e) 0 E V .
(2) L and M are maps from V to V satisfying: (a) (b) (c) (d) (el
Vx,y E V LxoLy < L ( x o y ) , L(a:ny) 5 L x n L y , 1 5 L 1 5 LT ( 5 T), vx, y E v Mx u My 5 M(x u y), (I5 ) M I 5 MO 5 0, VX,Y E V L(x+y) 5 Mx-+My , L(x+’y) 5 M x 4 M y .
To obtain FLKD-algebra,which corresponds to FLKD,we add to FLKalgebra the following condition (2)-(f) for operators L and M : (2)-(f) vx E
v Lx 5 Mx
We obtain the FLKT-algebra, which corresponds to FLKT, by adding to FLK-algebra the following condition (2)-(g), instead of (2)-(f): (2)-(g) VX E V LX 5 x
, z 5 MX
If we add the following condition (2)-(h) to FLKT-algebra, we obtain the FLKT4-algebra,which corresponds to FLKT4: (2)-(h) Vx E V Lx 5 LLx , MMx 5 Mx
432
To handle classical substructural modal logics, we introduce the basic algebra which is called CFLF-algebra. It is obtained by requiring the commutativity for the monoid mentioned in the condition (b) of the definition of FL-algebra, and furthermore by adding the “classical” condition (f) which is specified below. In classical cases, we do not need to distinguish +’ from + due to the commutativity of the monoid. Therefore, -+’is not necessary in the definition in CFLFalgebra, so that the last clause of (1)-(d) is deleted. Furthermore, L and M become mutually definable with the help of the “classical” negation. Thus, the conditions (2)-(c),(d) and (e) are to be discarded. Putting all these together, one comes down to the following definition of CFLF-algebra.
A structure A = (V,U, n, 0,+,L , M , 1,0, T, I)is a CFLE-algebra if the following conditions are satisfied: (1) (V,U, n, 0 , +,1,0, T , I)is a CFL,-algebra: (a) (V,U, n, T, I)is a lattice with the least element I and the greatest element T for which T = I + I holds, (b) (V,0,1) is a commutative monoid with the identity 1, (c) V x , y , z , w E V z o ( x u y ) o z u = ( z o x o z u ) u ( z o y o w ) , (d) V x , y , z ~ Vx o y < z @ x < y + z , (el 0 E (f) vx E v ( x c 0 )+ o = 2.
v,
(2) L is map from V to V satisfying: (a) v x , y ~ VL x o L y < L ( x o y ) , L ( x n y ) s L x n L y , (b) 1 5 L1 5 LT ( 5 T ) , To obtain CFLFD-algebra, which corresponds to CFLED, we add to CFLf-algebra the following condition (2)-(f) for operators L and M : (2)-(f) vx E
v Lx 5 L(x+O) + o
We obtain the CFLFT-algebra, which corresponds to CFLfT, by adding to CFLf-algebra the following condition (2)-(g), instead of (2)-(f): (2)-(g) vx E
v Lx 5 x
If we add the following condition (2)-(h) to CFLfT-algebra, we obtain the CFLET4-algebra, which corresponds to CFLFT4: (2)-(h) VX E V LX 5 LLx
433
We obtain the CFLFT5-algebra, which corresponds to CFLET5, by adding to CFLrT-algebra the following condition (2)-(i), instead of (2)(h): (2)-(i) Vx E V L ( x -+ 0) -+ 0 5 L ( L ( x-+ 0) -+ 0) Next, we define valuation map v : 7.=
+V.
Definition 3.2. (Valuation map) Let P be the set of propositional variables and let 210 be a mapping from P to V, then v is defined by the following recursion: (1) 4 P i > = voki) (2) v ( i A ) = w(A)+ 0 , v(1’A) = v ( A )+’0 (3) v(A > B ) = v(A) +v(B), v(A >’ B ) = v(A) +‘v(B) (4) v ( A * B ) = v ( A ) 0 w(B), v(A B ) = v ( i ( i A * -8)) ( 5 ) v ( A A B ) = v ( A ) n v ( B ) , v(A V B ) = v(A) U v ( B ) (6) v ( 0 A ) = Lv(A) , v(OA) = Mw(A) (7) v(t) = 1 , v(f) = 0 , v(T) = T , v ( l ) = I
+
In order to keep our exposition somewhat simpler, we will use a new parameter “L”. Let L be one of the following systems: CFLE, C F L r D , C F L f T , C F L r T 4 ,CFLtT5, FLK, FLKD,FLKT and FLKT4
Definition 3.3. (Validity of a sequent) A sequent I‘ + A is valid in L iff for any L-algebra and any valuation v we have v(r,) 5 .(A*), where l?* is the multiplicative conjunction of all the formulas in r and A* is the multiplicative disjunction of all the formulas in A. If r is the empty list of formulas, l? is treated as t. If A is empty, A is treated as f . Note: Classical modal logic K is characterized by the axiom O ( A > B ) > (CIA 3 O B ) . An alternative way to characterized K is to use the axiom O ( A A B ) OAAOB.This alternative method leads readily to an algebraic interpretations of non-substructural modal logics, as in [l],just by requiring L ( x n 3) = LXn Ly to hold. If one moves from non-substructural modal logics to substructural modal logics, then one can not always expect that O A A OB + O ( A r\ B ) is provable. In substructural logics, however, there are two kinds of logical conjunctions: Adaptation of one of them, i.e., multiplicative conjunction,
434
makes it possible to prove OA figure shows:
* O B -+ O(A * B ) , as the
A -+ A A , B -+ A * B UA
following proof
(* right)
* O B ?r D(A * B )
Based on these consideration, we have chosen the inequality Lx o L y L ( x o y) in the clause (2)-(a) of the definition 3.1.
5
4. Soundness
Theorem 4.1. (Algebraic soundness) If I‘ -+ A is provable in L, then r -+ A is valid in L.
Proof. It suffices to show that all the axioms are valid in L, and that every inference rule preserves validity, i.e., that for every inference rule of L, if the upper sequent(s) are valid, then the lower sequent is valid. The subsystem obtained from F L by deleting 3‘ has the “uni-residuated lattice-ordered groupoid”(0btained from FL-algebra by deleting -+‘) as its model, and the soundness of the subsystem is already proved in Lemma 2 of [2]. This Lemma can be applied not only to F L but also to CFL,. Therefore, we consider only the inference rules involving modality. ‘‘means j” that “from this (these) it follows that”. We use the following two simple facts. First, if y 5 y‘ then z o y o z 5 x o y‘ o z , because z o y‘ o z = z o (y U y’) o z = x o y o z U x o y‘ o z . Second, if x _< y then Lx 5 Ly, because Lx = L(z n y) _< Lx n L y 5 Ly.
(0-K) Assume that A l , . . . , A, -+ B is valid in L. Then, v(Al * . . . * A,) 5 v(B)is hold for any L-algebra. + v(UA1 * . * . * OA,) = Lv(A1) o . . . o Lv(A,) 5 L ( v ( A l ) o . . . o ~ ( A , ) ) = Lv(Ai*...*A,) 5 L v ( B ) = W ( U B ) Therefore, O A , , . .. ,UA, -+ O B is valid in L.
.
(O-K1) Assume that Al,. . . ,A,, B -+ C is valid in L (for intuitionistic cases). Then, v ( A l * . . . * A, * B ) 5 v(C) is hold for any L-algebra. + v(A1 * . . . * A,) 0 v ( B ) 5 w(C)
435
.
(V-K2) Assume that A, B1, . . . ,B, + C is valid in L (for intuitionistic cases). Then, v ( A* B1* . . . * B,) 5 v(C) is hold for any L-algebra. + v(A)0 v(B1 * . . . * B,) 5 w ( C ) + v(B1 * . . . * B,) 5 v(A)-+‘v ( C ) L w ( B ~.*. * * B,) 5 L ( v ( A )+’w ( C ) )5 M w ( A )+’ M v ( C ) Mv(A)0 L w ( B ~... * * B,) 5 M v ( C ) + v(OA * OBI * * * OB,) = M w ( A )o v(OB1 * . . . * OB,) 5 M I J ( A 0) L w ( B ~ * .* B,) 5 M w ( C )= w(OC) Therefore, OA, O B I , .. . , O B , + VC is valid in L.
+ +
(0-D) Assume that A1 , . . . , A, + is valid in C F L t D . Then, v(A1 * . . . * A,) 5 v(f) = 0 is hold for any CFLFD-algebra. v(A1 * . . . * A,-l) 0 v(A,) 5 0
+ + I J ( A ~* *. .* An-i) 5 v(An)+ O + Lv(A1 * . . . * An-l)
5 L(w(A,) -+ 0 ) 5 L((v(A,) -+ 0 ) + 0) + O = Lv(A,) + 0 + L z J ( A* .~. * An-1) 0 Lv(A,) 5 0 + v(OA1 * . . * OA,) = v(OA1 * . . . * CIA,-,) o v(OA,) 5 Lv(A1 * . . . * A,-I) o Lv(A,) 5 0 Therefore, OAl,. . . ,OA, + is valid in C F L r D . *
*
.
(UV-D) Assume that A1 , . . . ,A, -+ B is valid in FLKD. Then, v(A1 * . . . * A,) 5 v ( B ) is hold for any FLKD-algebra. + Lv(A1 * ... * A n )5 L v ( B ) 5 M v ( B ) + v(OA1 * . . . * DA,) 5 Lv(A1 * ... *A,) 5 M w ( B )= w ( V B ) Therefore, KIA,, ...,OA, + V B is valid in FLKD.
436
(0-T) Assume that A l , . . . , A , + B1,. . . , B , is valid in L (for CFLfT, CFLfT4, CFLFT5,FLKT and FLKT4). Then, v(Al *. . . *A,) 5 v(B1+. . . + B,) is hold for any L-algebra. j v(A1 * . * . * Ai-1 * DAi * Ai+l * * * . * A,) = v(A1 * . . * * Ai-1) 0 L v ( A ~0)v(Ai+l * . . . * A,) 5 v(A1 * . . * * Ai-1) o v(Ai)o ~ ( A i +*l . . * * A,) = v(A1 * . * * An) 5 v(B1 -I- . * . B,) Therefore, Al, . . . ,Ai-1, OA,, Ai+l,. . . , A , + B1,. . . ,B, is valid in L.
+
(0-T)
Assume that A l , . . . , A, + B is valid in L (for FLKT and FLKT4). Then, v(A1 * . . . * A,) 5 v(B)5 M v ( B ) is hold for any L-algebra. Therefore, A l , . . . , A , + O B is valid in L. 0
(0-4) Assume that OA,, . . . , OA, + B is valid in L (for CFLFT4 and FLKT4). Then, v(OA1 * . * OA,) 5 v ( B ) is hold for any L-algebra. jLv(OA1 * * . . * OA,) 5 L v ( B ) + v(OA1 * * * * OA,) = Lv(A1) o . . . O Lv(A,) 5 LLv(A1) 0.. . O LLw(A,) 5 L(Lv(A1)o . * * o Lv(An))= Lv(oA1 * . . . * DA,) < Lv(B)= v ( 0 B ) Therefore, OA,,. . . ,OA, + O B is valid in L. (0-41) Assume that CIA,, . . . , OA,,B + OC is valid in FLKT4. * OA, * B ) 5 v(0C) is hold for any FLKT4Then, v(OA1 * algebra. + v(OA1 * . . * * DA,) 0 v ( B ) _< M v ( C ) + v ( o A l * * * . * O A , )~ U ( B ) + M V ( C ) + Lv(OA~***.*OA,) 5 L(v(B)+Mv(C))5 M w ( B ) + M M v ( C ) + Lv(oAi* * . * * =A,) o M v ( B ) 5 M M v ( C ) 5 M v ( C ) 3 v(OAi* . - * * CIA, * OB) = v(oA1* . * * * OA,) 0 M v ( B ) 5 Lv(oAi* . . . * OA,) 0 M w ( B )5 M v ( C ) = v ( 0 C ) Therefore, OAl,. . . , DA,, Q B OC iis valid in FLKT4.
437
(0-42) Assume that A, O B I , .. . ,UB, -+ OC is valid in FLKT4. Then, v(A*UB1*.. .*LIB,) 5 v ( 0 C ) is hold for any FLKT4-algebra. =+ v ( A ) 0 v(OB1* . . . * OB,) 5 M w ( C ) jv(oB1 * * . . t OB,) 5 v ( A )-+'M v ( C ) jLv(OB1* . * . * OB,) < L(w(A)-+'M v ( C ) ) 5 Mv(A)-+'M M v ( C ) jMv(A) 0 Lv(oB1 * . * .* OB,) 5 M M w ( C )5 M w ( C )
+v(OA*OB1 * . . . * OB,)=MV(A)O~(OB~*...*OB,) 5 Mv(A)o L v ( O B ~ **.. * LIB,) 5 M w ( C )= v ( 0 C ) Therefore, OA, O B I , .. . , U B , -+ OC is valid in FLKT4. 0
(0-5) We use the following facts which hold for CFLzT5-algebra. ( L ( L x+ 0) -+ 0 ) 5 Lx holds, because: L ( ( x-+ 0) + 0) + 0 5 L ( L ( ( x-+ 0) 4 0) -+ 0 ) =+ LX-+O5 L(Lx-+O) = (L(L2+0)+0)+0 =+ (LZ-kO) 0 (L(Lx-+O)+O) 5 0 (L(Lx-+O)-+O)5 (Lx+O)-+O = Lx x 5 L ( z 0) -+ 0 holds, because: L(2 3 0) 5 (x-+ 0 ) =+ x 5 L ( x -+ 0 ) -+ 0 Lx 5 LLx holds, because: L x ~ L ( L z t O ) - , O ~ L ( L ( L x - + O ) - +5OL)L x is valid in Assume that OA,,. . . , OA, + B , OC1,. . . , OC, CFLfT5. holds for Then, v(OAl * - .. * OA,) 5 v ( B OC1 . . . UC), CFLET5-algebra. j v(OA~*****OA 5, (v(B)+O)o(v(OC1 ) +...+OC,)+O)-+O jv ( O A i * . . .* OA,) o ( v ( B )4 0 ) o ( ~ ( 0 C i ...+ + OC,) + O ) 5 0 jv(oA1 * . . . * OA,) o ( w ( B-+ ) 0) < (v(OC1+ . .. OC,)+O) 4 0 = v(OC1 * .. OC,)
*
+
+
+
+
+ +
438
5. Completeness Definition 5.1. (The Lindenbaum algebra) We define IAl to be the set { B : A + B and B + A are provable in L}. Each J A Jconstitute an equivalent class, and it does not depend on the choice of the representable element. Among these equivalent classes, we introduce algebraic operations in the following way:
IAl o IBI = [ A* BI J A ( r l [ B I = [ A A B ,J( A ( U I B [ = ( A V B ( \A(+ IBI = \A 3 B ( , (A1+‘ IB( = ( A3’B ( LlAl = IoAl , MIA1 = 10-41 Let F*= {IAJ : A E 3). The Lindenbaum algebra of L is defined to be the algebra:
v*, n, u,
+’, L , M , Itl, Ifl, ITI, 14)
0 ,+>
(For classical cases, +’ and M are excluded.)
L e m m a 5.1. The Lindenbaum algebra of L is an L-algebra. Proof. As in the proof of Theorem 4.1, we check here only the conditions involving modality. (For the other conditions, we refer the reader to Lemma 3 of [2].) Take A and B to be any formulas.
439 We need to show the fact that (A1 5 IBI is hold in the Lindenbaum algebra of L if A + B is provable in L. Assume that A + B is provable in L. Then, we have the proof
A
A A -t B (A left) A A (A right) A-+AAB AAB-+A
+
+
From this, it follows that JAl= IAABI = IAl n IBI holds in the Lindenbaum algebra of L. Hence, IAl 5 (BI holds. Using the above fact, we can show that the Lindenbaum algebra of L satisfies the conditions of L-algebra in the following ways.
5 L(IAI o IBI) holds, because we have
0
LlAl
0
L(IAI n IBI) 5 LIA( n LIB1 holds, because we have
o
LIB1
A + A (A left) A A B B (A left) AAB+A O(A A B ) -+ O B (0-W ~ ( ABA) + UA (A right) U(AA B ) + OA A UB -+
+
It1 5 Lltl 5 LIT1 5 IT1 holds, because we have
L(IAI + IBI) 5 MIA1 + MIBI L(lAI +’ PI) holds for intuitionistic cases. because we have 7
I
MIA1 -+’MlBl
440
(I( 5 MIL( 5 M(f(5 If1 holds for intuitionistic cases, because we have f + I + f
01+ Of (0-K1) Of
I + Ol(axiom) 0
LIA( _< L((AI+ lfl)
+ If(
-$
(O-K1) (fw)
holds for CFLFD, because we have
L/AI 5 MIA1 holds for FLKD,because we have A + - A (00-D) UA + OA 0
L J A J5 IAJ holds for CFLrT, CFLfT4, CFLFT5, FLKT and FLKT4,because we have
*
A A (0-T) OA + A 0
IAJ 5 MIA1 holds for FLKT,FLKT4,because we have A+A A+OA
(0-T)
0
L J A J5 LLJAJholds for CFLfT4 and FLKT4,because we have
0
IMMJAI5 MIA( holds for FLKT4,because we have
441
0
L()AI+If[) + If) 5 L(L(IAI+ If[)+ If[) holds for CFLFT5,because we have
Theorem 5.1. (Algebraic completeness) If a sequent I' + A is valid in L, then I? + A is provable in L. Proof. Suppose that A -+ B is valid in L. Then, IAl 5 lBl holds in the Lindenbaum algebra of L. From this, it follows that )A1 = ) A )n ) B )= ) A A B ) . By the definition of equivalence classes, A + A A B is provable in L. Now then, we can construct a following proof of A + B in L.
'
( A left) A-+AAB A A B + B (cut) A+B +
The above argument for A
+ B carries over for I? -+ A.
6. Conclusion
We have extended the sequent systems FL and CFL,, which constitute the basis of substructural logics, to obtain nine new systems for substructural modal logics CFL:, CFLFD,CFLFT,CFLfT4, CFL5T5,FLK, FLKD, FLKTand FLKT4by the addition of inference rules for the modal operators 0 and 0.These systems are proved to be both sound and complete through algebraic interpretations. Non-substructural modal logics have been recently a target of extensive research activities motivated by not only theoretical interests, but also various kinds of intentions to apply them to fields related to foundations of computer science, AI, and control engineering, to mention a few. We think that substructuralization of these modal logics will make it possible to obtain more detailed and finer arguments and analysis in these fields than those obtained by merely applying non-substructural modal logics. In this
442
regards, our soundness and completeness theorems may be regarded as providing a methodological foundation t o future applications of substructural modal logics. References 1. B. Chellas. Modal Logic: an introduction. Cambridge University Press, Cambridge, UK, 1980. 2. K. Dogen. Sequent systems and groupoid models. I. Studia Logical 47:353-385, 1988. 3. M. Ohnishi and K. Matsumoto. Gentzen method in modal calculi. Osaka Mathematical Journal, 9:113-130, 1957. 4. H. Ono. Semantics for substructural logics. In K. Dogen and P. ShroederHeister, editors, Substructural Logics, pages 259-291. Oxford University Press, Oxford, UK, 1993. 5. H. Ono. Proof-theoretic methods in nonclassical logic - an introduction. In Theories of Types and Proofs, MSJ Memoirs, pages 207-254. Mathematical Society of Japan, Tokyo, Japan, 1999. 6. A. Simpson. The Proof Theory and Semantics of Intuitionistic Modal Logic. PhD thesis, University of Edinburgh, 1994. 7. S. Valentini. The sequent calculus for the modal logic D. Bollettino della Unione Matematica Italiana, 7-A:455-460, 1993.
443
DIAMOND EMBEDDINGS INTO THE D.C.E. DEGREES WITH 0 AND 1 PRESERVED
GUOHUA WU School of Mathematical and Computing Sciences Victoria University of Wellington P.O. Box 600, Wellington New Zealand email: wu(0mcs. vuw. ac .nz
1. Introduction
Say that a set A C w is computably enumerable (c.e. for short), if A can be listed effectively. Thus we can define a set D C_ w to be d.c.e. if D is the difference of two c.e. sets. A Turing degree is c.e. (d.c.e.) if it contains a c.e. (d.c.e.) set. Let R be the set of all c.e. degrees and D2 be the set of all d.c.e. degrees. Since any c.e. set is d.c.e., R C_ Dz. Cooper [4] showed that there are d.c.e. degrees containing no c.e. sets (these d.c.e. degrees are called properly d.c.e. degrees), and hence R c D2. In [6], Cooper, Lempp and Watson proved that the properly d.c.e. degrees are densely distributed in the c.e. degrees. Lachlan observed that any nonzero d.c.e. degree bounds a nonzero c.e. degree, and so the downwards density holds in D2. Thus as noticed by Jockusch, Dz is not complemented. The first two structural differences between Dz and R are:
Theorem 1 (Arslanov’s Cupping Theorem [2]) Every nonzero d.c.e. degree cups to 0’ with an incomplete d.c.e. degree. Theorem 2 (Downey’s Diamond Embedding Theorem [lo]) There are two d.c.e. degrees dl, d2 such that dl U d2 = 0’, dl n d2 = 0.
By Theorem 2, the diamond lattice can be embedded into the d.c.e. degrees preserving 0 and 1. In this paper, we will refer to such embeddings as Downey ’s d i a m o n d embeddings. In [16], Li and Yi constructed two d.c.e. degrees dl,d2 such that any nonzero c.e. degree cups one of dl, d2 to 0’. As a corollary, any nonzero d.c.e. degree below dl is a complement of dz, and hence, one of the atoms
444
in Downey’s diamond can be c.e., which was demonstrated first by Ding and Qian in [9]. In [5], Cooper et al. proved that D2 is not densely ordered, which gives a more striking difference between R and D2: Theorem 3 (Cooper, Harrington, Lachlan, Lempp and Soare [5]) There
is a maximal incomplete d.c.e. degree. By Lachlan’s observation, the dual version of Theorem 3 is not true. However, as proved by Cooper and Yi [7], and independently by Ishmukhametov [12], the following weak density holds in the d.c.e. degrees: Theorem 4 (Cooper and Yi [7], Ishmukhametov [12]) For any c.e. degree a and d.c.e. degree d, if a < d, then there is a d.c.e. degree e such that
a<e
It is easy to see that d is isolated by a if and only if the interval (a,d) contains no c.e. degree. The isolated and the nonisolated d.c.e. degrees are densely distributed in the c.e. degrees (see Ding and Qian [8], LaForte [15], and Arslanov, Lempp and Shore [3]). In 1131, Ishmukhametov and Wu proved that there is an isolation pair (a,d) such that a is low and d is high. Thus, in the sense of the highflow hierarchy, the isolated degree can be far from the isolating degree. This paper is mainly concerned with the relationship between Downey’s diamond embeddings and the isolation phenomenon. First, we have the following observation: Proposition 6 (Wu [19]) For any c.e. degree c and isolation pair (a,d), if c n a = 0, then c n d = 0.
Proof of Proposition 6: Suppose not. Let c be any nonzero c.e. degree. Then by Lachlan’s observation that any nonzero d.c.e. degree bounds a nonzero c.e. degree, there is a c.e. degree e E (0,c n d). Thus, e < d and hence e < a since a bounds all c.e. degrees below d. an c 2 e > 0. A contradiction. 0 Based on this observation, Wu [19] gives an alternative proof of the existence of Downey’s diamond embedding:
445
Theorem 7 (Wu [19]) There are an isolation pair (a,d) and a c.e. degree c such that c u d = 0‘, c n a = 0. Thus, (0, c, d, 0’} is a Downey’s diamond embedding. The main idea involved in Theorem 7 is to ask d to be responsible for cupping c t o 0’ and a to be responsible for capping c to 0. This appears to be as close as we can get to overcoming the obstacle to constructing the diamond embedding into the c.e. degrees preserving 0 and 1, which first became apparent in Lachlan’s Non-Diamond Theorem [14]. Furthermore, Wu [22] showed that this idea can also be used to prove that for any high c.e. degree h, the nondistributive lattice M:, can be embedded into the interval [0,h] preserving 0 and 1. As a consequence, the nondistributive lattice S8 can be embedded into the d.c.e. degrees with 0 and 1 preserved. Say that a c.e. degree c is cappable in R if there is a nonzero c.e. degree b such that c n b = 0. Let M be the set of all cappable degrees. Obviously, 0 E M. Let NC = R - M. Ambos-Spies et al. [l]showed that M is an ideal in R and N C is a strong filter in R. They also showed that a c.e. degree c is cappable if and only if no low c.e. degree can cup c to 0‘ if and only if c contains no promptly simple sets. Downey, Li and Wu [ll]gives a new characterization of cappable degrees: c is cappable in R if and only if c has a nonzero complement in the d.c.e. degrees.
Theorem 8 (Downey, Li and Wu [ll]) For any c.e. degree c > 0, c is cappable in R if and only if there is an isolated degree d such that cud = 0’, cnd=O. Theorem 8 says that any nonzero cappable degree can always have an isolated degree as its complement. Again, the proof of Theorem 8 involves a construction of an isolation pair (a,d), such that c caps a to 0 and cups d to 0’. In this paper, we prove that any nonzero cappable degree can always have a nonisolated degree as its complement.
Theorem 9 For any c.e. degree c > 0, c is cappable in R if and only if c can be complemented in the d.c.e. degrees by a nonisolated degree. To prove Theorem 9, it’s necessary to prove:
Theorem 10 For any c.e. degree c > 0, if c is cappable in R, then there are two d.c.e. degrees b < d such that (1) b is properly d.c.e. and nonisolated; (2) b n c = 0, d U c = 0’; (3) b bounds all c.e. degrees below d.
446
Proof of Theorem 9: First, suppose that c has a complement d in the d.c.e. degree. Then by Lachlan's observation, we know that there is a nonzero c.e. degree e below d. c n e = 0. c is cappable in R. Now fix c > 0 as a degree cappable in R. By Theorem 10 (2) and (3), c is complemented by d. Now we show that d is nonisolated. Let e be any c.e. degree below d. Then by the fact that b bounds all c.e. degrees below d, we have e < b. Since b is nonisolated, there is some c.e. degree O f between e and b, and hence between e and d. d is nonisolated.
(b,d) in Theorem 10 is called a pseudo-isolation pair. For more details on the pseudo-isolated degrees, see Wu [HI, [20], [21]. We organize the paper as follows. In section 2, we list the requirements needed to prove Theorem 10, and describe the basic strategies to satisfy these requirements. In section 3, we give the construction, and in section 4, we verify that our construction satisfies all the requirements. Our notation and terminology are quite standard. During the construction, when we define a parameter as a fresh number x at stage s, we mean that z is the least number greater than any number mentioned so far. Particularly, x > s. For others, see Soare [17]. 2. Requirements and basic strategies Given a cappable c.e. degree c > 0, let C E c be any c.e. set. To prove Theorem 10, we will construct two d.c.e. sets B , D ,an auxiliary c.e. set E , and a partial functional satisfying the following requirements:
6: K = r(C, D); PE: E # P,": D # @ ;: Me: B # @,W. V We # iP:; N, : @: = @f= g total + g computable; R, : +pD = we+ ~ A , ( A : = we); Q, : W, = 0:
+ (3c.e. U, ST B ) ( v ~ ) (#u ,OF).
(2-1) (2.2) (2.3) (2.4) (2.5) (2.6) (2.7)
where e E w , {(@,,We) : e E w } is an effective enumeration of all pairs (a, W ) such that cf, is a partial computable functional, W is a computably enumerable set. K is a fixed creative set. Let b , d , e be the Turing degrees of B , B CBD, E respectively. By the requirement G, c U d = 0'. By the N-requirements, c n b = 0 . By the M-requirements, b is a properly d.c.e. degree. By the PE-requirements, d is incomplete. Thus, c and d are incomparable.
447
We now show that d is pseudo-isolated by b. By the PD-requirements, b < d. The Q-requirements guarantees that b is nonisolated and the Rrequirements guarantees that b bounds all c.e. degrees below d.
2.1. The 8 - s t r a t e g y
In the construction, the G-strategy will be responsible for coding K into C @ D . The G-strategy proceeds as follows: If there is an z such that r(C, D ; z ) [ s ]$# K,(x), then let k be the least such z, enumerate y(lc)[s] into D , and for any 1~ 2 Ic, let r(C, D ; y) be undefined. Otherwise, let k be the least number z with r(C, D; z)[s] f. If r(C, D ; z) has never been defined so far, then set r(C, D ; k ) [ s ]= K,(k) with y(k)[s] fresh. If not, let t be the last stage at which r ( C , D ; k ) [ t ]4. If one of the following holds, then set r(C, D; k)[s] = K,(k) with y(k)[s] fresh.
(a) There is some y < k with y(y)[s] > y(k)[t]; (b) There are some y-markers less than or equal to y(Ic)[t]enumerated into D or removed from D after stage t;
(In the construction, if a y-marker z is enumerated into D at stage and moved out at stage s2 > s1, then between these two stages, C will have a change below z, which allows us to list this y-marker to a larger number.) (c) C has a change below an active requesting number (as defined later), z say, and z 5 y ( k ) [ t ] . s1
If (a)-(.) do not apply, then set r(C, D ; k ) [ s ]= K,(k) with y ( k ) [ s ]= y ( k ) [ t ] . The G-strategy guarantees that r(C,D ) is totaIly defined and computes K correctly. Obviously, y-markers have the following properties:
448
2.2. A PE-strategy
A PE-strategy, a say, will satisfy a PE-requirement, E #
say. a is the Friedberg-Muchnik strategy, with some modification to cooperate with the coding of K . That is, during the construction, the coding procedure may enumerate infinitely many numbers into D and hence may injure the standard Friedberg-Muchnik strategy infinitely often. To avoid this, we use the threshold strategy as follows. Set a parameter k ( a ) first as a fresh number. k ( a ) acts as a threshold for the enumeration of y-markers. Whenever K changes below k ( a ) , reset a by canceling all parameters of a , except k ( a ) . Since k ( a ) is fixed, such a reset procedure can happen at most finitely many times. Let SO be the last stage at which a is reset or initialized. Suppose that at some stage s l , @:@D(z(a))[sl] converges to 0, then instead of putting .(a) into E immediately, we put y ( k ( a ) ) [ s l into ] D first to lift y(z) for z 2 k ( a ) to big numbers, and request that y ( k ( a ) )be undefined whenever C has a change below y ( k ( a ) ) [ s l ] .Correspondingly, we call y ( k ( a ) ) [ s l ]a requesting number. Say that y(k(a))[s1]’srequest is realized if C has a change below y ( k ( a ) ) [ s l and ] that y ( k ( a ) ) [ s 1 ] % request remains active if a has not been initialized or reset or y ( k ( a ) ) [ s l ] ’request s has not been realized. Note that the enumeration of y ( k ( a ) ) [ s linto ] D prevents the G-strategy from now on. However, from injuring the computation Qi:@D(z(a))[sl] ] into D may injure the computation the enumeration of y ( k ( a ) ) [ s l itself + : @ D ( z ( a ) ) [ ~Such l ] . injuries are called the “capricious injury” , which was first used by Lachlan in his nonsplitting theorem. Now suppose that y ( k ( a ) ) [ s l ] ’ srequest is realized at stage $ 2 , i.e., Cszr y ( k ( a ) ) [ s l # ] C,, y ( k ( a ) ) [ s l ] .Then y ( k ( a ) ) [ s 2 ]is redefined as a big number as requested by Y ( k ( 4 ) b l I (particularly, Y(k(Q))[S21> r(k(a)>[s11),and r ( k ( a ) ) [ s 1 I 1 s request becomes inactive forever. Let s3 2 s2 be the next a-stage. Then, by taking y ( k ( a ) ) [ s l out ] of D , the computation 4jf@D(z(a)) is recovered [ ~ ~y]( k, ( a ) )is undefined again by D,, ty(lc(a))[sz]# to @ : @ D ( z ( a ) )and Dsz t y ( k ( a ) ) [ s 2 ] Furthermore, . in the construction, to cooperate with the R-strategies (see below), at stage s3, when we take the number y ( k ( a ) ) [ s l ] out of D , we also put s1 into B. The enumeration of s1 does not injure the PE-strategy described above because s1 > cp,(z(a))[sl]. We now describe the a-strategy in details. First, choose z ( a )as a fresh number. Particularly, .(a) > k ( a ) . a runs cycles ( n ) n, E w . Fix n. Cycle (n)runs as follows: (1) Wait for a stage sn at which @:@D(z(a))[~n] $= 0.
449
(2) Put r ( k ( a ) ) [ s ninto ] D. For those z < r(k(a))[s,]with g m ( z ) f - , define g a ( z ) = Cs,(z). Declare that y(k(a))[s,] requests that C have a change below it to perform the Friedberg-Muchnik strategy. Start cycle ( n l),and simultaneously, wait for C to change below Y ( ~ ( Q ) [) ~ n l . (3) Enumerate .(a) into E , and simultaneously, take r ( k ( a ) ) [ s nout ] of D ,enumerate sn into B. Stop.
+
Since C is noncomputable, g , cannot be totally defined. That is, not every cycle can reach (2) and wait at (2) permanently. Let (n)be the first cycle reaching (3). Then a is satisfied because E ( x ( ~= ) )1 #
o = @feD(z(a))[~nl = +feD(~(a)).
a has the following outcomes:
0
."
where n means that the cycle ( n ) waits at (1) forever, and d means that some cycle reaches (3) eventually. 2.3. A IPD-strategy
A PD-strategy, p, is a standard Friedberg-Muchnik strategy. follows:
p works as
(1) Define x(P) as fresh. (2) Wait for a stage s at which + f ( z ( p ) ) [ s ] $= 0. (3) Enumerate z(@)into D and stop. ,O has two possible outcomes, 0 , l with 0 < L 1, where 0 indicates that ,B arrives at step (3) eventually, and 1 indicates that a keeps waiting at step
(2). 2.4. A n M - s t r a t e g y
An M-strategy, q say, working for an M-requirement, Me, attempts to find a number witnessing B # @? or We # Q:.
q works as follows:
(1) Choose ~ ( qas) a fresh number, (2) Wait for a stage s such that 0 = Bs(z(q))= +?(X(~))[S]
Qt:
and
t ve,s(z(q)). We,, t pe,s(z(q)) = (If this never happens, then z ( q ) is a witness to the success of Me.)
450
(3) Put z(q) into B. Protect B, r s from other strategies. (4) Wait for a stage s‘ such that 1 = B,, ( z ( q ) )= a? (x(q))[s’]and
We,sl tpe,s(x(q)) =
tpe,s(x(q))-
(If this never happens, then again x ( q ) is a witness to the success of M e . If it happens, then the change in (x(q)) between stages s and s‘ can only be brought about b y a change in We rpe,s(x(q)), which is irreversible since We is a c.e. set.) (5) Remove x(q) from B and restrain B r s from other strategies. (Now x ( q ) is a permanent witness to the success of M e because We tpe,s(x(q)) # tpe,s(x(q)).)
*:
q has two outcomes, 0
To satisfy an N-requirement, N, say, we will define (partial) computable functions q!Je,i,i E w such that for some i, q!Je,i = he, provided that = = he is total. Define the length of agreement between @fand @: as follows:
[(e,s) = .{ :VY < x[@,B(d[s1-1=@,c(Y)[S14h m ( e , s) = max{l(e, t ) : t < s}. Say that stage s is expansionary if s = 0 or [ ( e ,s) > m ( e ,s). Our strategy is to preserve computations up to the greatest length of agreement between @: and a:. Since C has a cappable degree, we can utilize the gap-cogap argument to preserve the computations. That is, we will define a partial computable function p , such that either N,-requirement is satisfied or p e witnesses that C has a promptly simple degree, which is impossible by our assumption on C. In the latter case, p , satisfies the following requirements: Se,i:Wi infinite
+
( 3 ~ ) ( 3 ~ )E [ aWi,a+, : ,& C, 1%# Cpe(,)1x1.
The whole construction is partitioned into infinitely many intervals which are referred to as gaps and cogaps alternatively (as defined later), and we only allow B to change at those stages inside a gap. Thus, inside a gap, we allow B to change and see whether C has a change on small numbers. If C does have a change on some small numbers inside a gap, then some computation @:(y) may change, and at the next expansionary
451
stage, both @F(y)and @s(y) converge to a new value, making +,,i(y) incorrect. However, such C-changes provide us opportunities to satisfy the &+-requirement if we define p,(s) properly (as the stage at which the gap is closed) and we can start to attack a new S-requirement. On the other hand, if C has no such changes during a gap, then at the stage we close the gap (another expansionary stage), we will set up a restraint on B to preserve the B-side computations. Thus, if @: = @: = he is total, then by our assumption that C has a cappable degree, there is some (least) i such that Se,i is not satisfied. In this case, Se,i will define +)e,itotally. As described above, inside the gaps, the C-side computations don’t change and inside of the cogaps, the B-side computations are preserved. Therefore, &?,i computes he correctly. Now we define gaps and cogaps formally. Let s be any expansionary stage. If there is some i such that
(01) Se,i is not satisfied; (02) x enters Wi at stage s; ( 0 3 ) 3Y[Y < q e , s) @ e , i ( Y ) [ S ] t & x
> max{cp(Cs; el Y’,3)
: Y’
I Yll.
Then let i be the least one and open a gap for Se,i as follows:
(1) For those y‘ _< y, where y is as in (03), if ?Jle,i(y’)[s]f, then define @e,i(Y’)= @?(Y’)[S]; (2) Set the restraint r ( e ,i, s) = 0, and initialize all strategies Se,j with j
> i.
Let v be the next expansionary stage and close the gap as follows: (Cl) Define p e ( s ) = v . (C2) Set r ( e ,i, v ) = v . During a gap, if C has a change below x,then by p e ( S ) = v , C, tx # Cpe(,)7 x, Se,i is satisfied. In this case, we say that the gap is closed successfully. If there is no such C-change, then the gap is closed unsuccessfully. Suppose that @: = @: = he is total. Since C has a cappable degree, there is a least i such that Se,icannot be satisfied. Then Se,iopens infinitely many gaps, and each one is closed unsuccessfully. Let so
< . . . < s,
< ...
be the stages at which Se,i opens and closes gaps alternatively. We prove below that I / . J ~computes ,~ he correctly.
452
Fix y and let $ ~ ~ , i ( ybe) defined at stage s,. at stage s,. That is,
Then we open an &+-gap
(1) sj is an expansionary stage, (2) some z enters Wi at stage s,,
and there is some y with (3) Y < t ( e , s n ) ,$ e , i ( ~ ) [ ~ Tand n] 2 > (~(Cs,;e,~,sn), (4) q!~~,i(y)is defined as +F(y)[s,] at the end of stage s,.
Then at stage w,, we close this gap by defining p e ( s , ) = w,, restraining numbers less than w, from entering B till stage s,+1. Since C has no change below cp(CSn ; e, y, s,) inside this gap (otherwise, Se,i will be satisfied), @F(y)[s,] = @F(y)[wn], and hence $e,i(Y)
= @ f ( y > [ ~ n= ]
+.S(Y>[S,I = +F(Y)[v,I = +?(Y>["~I.
Now numbers less than w, are restrained from entering B between stages w, and s,+1, numbers less than w, are restrained from entering B , and as a result, the computation +:(y)[w,] is preserved and hence +e,i(Y)
= + f ( ~ > [ w n I= + f ( ~ ) [ ~ n + l= I +?(y)[~n+lI.
By induction, we have for all m 2 n, $e,i(Y)
= @ f ( ~ ) [ ~ r n=I +?(y)[~mI = +?(~)[wrnI = + f ( ~ > [ w r n ] .
Since both @(y) and +F(y) converge, we have $e,i(Y)
= +:(Y)
= +?(Y> = h e ( Y ) .
Let a be any N-strategy on the tree. Then rs works to satisfy requirement. Define
Ne(,)-
l ( o , s )= {z : VY < "[@$,)(Y)[sl-1=+~,)(Y)"}; m(a,s) = max{l(a, t ) : t < s and t is a-stage}. Say that a stage s is a-expansionary if s = 0 or s is an a-stage and qrs, s) > m(a,s). rs has infinitely many substrategies, each of which works on an Se(u),irequirements. In the following, we write S,,i for Se(,,),i,$,,i for $e(u),i for convenience. During the construction, S,,i may open (and hence close) gaps at expansionary stages, and whenever S,,i opens a gap, S,,i will extend the definition of qm,i. Say that S,,i requires attention at an a-expansionary stage s if one of the following holds:
(1) S,,i is inside a gap. (2) S,,i is inside a cogap. There are two subcases:
(2A) There is some y E dom($,,i) such that C, cp,(C; y)[v] # C, cp,(C; y)[v], where v is the last a-expansionary stage. (2B) &,i is ready to open a gap.
r
In case (2), (2A) has higher priority than (2B). It may happen that (2A) prevents S,,i from opening a gap (2B) for almost all times. In this case, dom(&,i) is finite, and there is some y E dom($,,i) with @F(g)T. Sg,i has two outcomes g
WYH. Say that S,,i receives attention at an a-expansionary stage s as follows if Sa,irequires attention at this stage:
Case I: (1)happens. Then close the gap, define p,(v) = s, and initialize all nodes with lower priority. Stop stage s. If C has a change below x (x is defined in (02) at stage w , where w is the stage at which the gap is opened) then we say the gap is closed successfully, and declare that S,,i is satisfied. Otherwise. The gap is closed unsuccessfully.
Case 2 (2A) happens. Then S,,i has outcome d. $,,i
Case 3: (2B) happens. Then S,,i opens a gap, extends the definition of according to (03). S,,i has outcome g.
In the construction, we don’t put S strategies on the priority tree. We just attach the outcomes of S to a. Thus, a has outcomes go
. . .
gi < L di
. ..
which are described as follows:
f. f denotes the case in which there are only finitely many
u-
expansionary stages (and hence, u is satisfied trivially). d . d denotes the case in which there are infinitely many aexpansionary stages, and a’s substrategies can require attention only finitely many times (if so,then is not total because (03) fails for almost all times and hence there is some z such that @ : ( z ) t , (a: is not total.)
454
gi. gi denotes the case in which the substrategy Su,i opens (and closes) gaps infinitely often. As described above, Gu,i is totally defined. di. di denotes the case in which Sg,ican open gaps only finitely often and there is some y E dom($,,i) such that @z(y) diverges. 2.6. A n U - s t r a t e g y
An %?,-strategy works to satisfy '&-requirement. Define l ( e , s ) = max{z < s : (Vy < z)(@:@D(y)[sI &=We,s(y))}, m(e, s ) = max{O,l(e, t ) : t < s} Say that stage s is expansionary, if s = 0 or l ( e ,s ) > m(e,s). The basic isolation strategy works as follows. At expansionary stage s > 0, for any y < t ( e , s ) , if Af(y)[s] t, then define Af(y) = W , , s ( y ) with 6,(y)[s] = s. Suppose that 6,(y) is defined at stage s. During the construction, numbers enumerated into D ,z say, may injure the current computation of @:@D(y). Such an enumeration may lift the use (pFeD(y), but with Af(y) unchanged. This provides chances for We t o change at y , and if so, A, becomes incorrect at y. Suppose that y enters We at stage s' > s. At the next expansionary stage s" 2 s', by taking numbers, z mentioned above, out of D ,we recover the computation @:eD(y) to @:@D(y)[s]. This creates an inequality because
By preserving this inequality, due to the c.e.-ness of W e ,y remains in W e , and therefore R, is satisfied. We refer to this method as the disagreement strategy. On the other hand, if is total and = W e ,then we will ensure that A: is total and computes W, correctly. Let J be any R-strategy on the tree. Instead of using a,(<), We(<) explicitly, we use at, W
t(t,s ) = max{a: < s : (VY < z)(@PeD(Y)[SI-1= m(J, s) = max(O,l(J, t ) : t
Wg,s(?/))},
< s and t is a J-stage}.
Say that stage s is (-expansionary, if s = 0 or s is a J-stage and l ( Js) , > m(t7 s). Similar to the PE-strategy, J needs to cooperate with the G-strategy. Set k ( J ) as a fresh number. k ( J ) will act as a threshold for the enumeration of y-markers and whenever K changes below k(J), reset t. Again, since k ( J ) is a fixed number, we can assume that after stage so,K has no change below k(J) and hence J cannot be reset afterwards.
455
Without loss of generality, suppose that at a (-expansionary stage s1 > E defines AF(y) = W C , ~ ~with ( ~ )use S,(y) = s1, and that at a (expansionary stage s 2 > s1, 6 finds that A? has a wrong guess of W,(y) (i.e., AF(y) = W C , ~ ~=( 0~ #) 1 = W€,sz(y)),then 5: will perform the following actions: SO,
(1) initialize all strategies with Iower priority to prevent these strategies from injuring the associated computation; ( 2 ) enumerate y(k(())[s2] into D to lift y(z) with z 2 k(() to prevent the G-strategy from injuring the associated computation. Again, we say that y(k(J))[s2] requests that C have a change below it to perform the disagreement strategy. Note that the enumeration of y(k(())[s2] may injure the associated computation. However, if r(lc(())[s~]'s request is realized (C ry(k(())[s~]changes) at stage s3 > s2 say, we redefine y(k(J))[s3] as a big number. Particularly, y(k([))[s3] > s3. Now at the next &expansionary stage, s4 2 s3 say, we take the numbers enumerated into D after stage s1, including y(k(c))[s2], out of D ,and thus recover the computation @FBD(y)to @FeD(y)[s1], which equals 0. By preserving this computation, we have BBD
@,BBD(Y)= a<
(Y"l1
= W€,S1(Y)= 0 # 1 = Wr,dY) = W Y ) .
R is satisfied. As in the PE-strategy, for the sake of the consistency between ( and those R-strategies with higher priority, at stage s4, we also put s1 into B , and declare that for all R-strategies E' c and for any y, if 6<1(y) > s1, let AF(y) be undefined. Obviously, the enumeration of s1 does not injure the disagreement created above. Now we describe the (-strategy in detail. runs cycles ( n ) ,n E w , each of which defines a partial computable functional, A<,n.
<
Cycle (n)runs as follows: (1) Wait for a (-expansionary stage. (2) Let s1 be a &expansionary stage. Then, for any y < t(6,sl) with A&(y)[sll t, define A&(Y)[sI] = W E , ~ ~ (with Y ) ~<(Y)[SI] = si. Go back to (1) and simultaneously, wait for W,(y) to change. (3) Let s 2 > s1 be the first (-expansionary stage with Wr,,,(y) = 1 # 0 = W ~ , s l ( ~and ) A&(Y)[S~]-1= A~,(Y)[sI] with use 6<(y)[s2]= s1. Put y(k(())[s2] into D ,and declare that y(k(<))[s2] requests that C have a change below it to perform the disagreement strategy. For those z < y(lc(())[52], if gc(z) t,define gc(z) = C,, ( z ) .
456
+
Start cycle ( n l), and simultaneously, wait for C to change below y ( k ( 5 ) )b 2 1 . (4) Take y ( k ( [ ) ) [ s aout ] of D ,put s1 into B (perform the disagreement strategy f o r 5) and declare that is satisfied via the disagreement at y .
<
( satisfies the R-requirement as follows:
Case 1. There is a cycle waiting forever at (1). (There i s some y such that @FeD(y)T or @FeD(y)$# W c ( y ) . ) Case 2. There is a cycle, ( n )say, going from (2) to (1) infinitely often. = Wc, then A:, is totally defined, and computes Wc (If correctly.) Case 3. There is a cycle reaching (4) via number y.
(Then, @F'D(y)$= 0 # 1 = W c ( y ) . ) Note that the incomputability of C ensures that it cannot be the case that for each n E w , cycle ( n ) is started, because otherwise, each cycle would wait at (3) forever, making gc totally defined, and hence C is computable. Thus, ( has the following outcomes:
( 0 , ~
Remark: Let A & ( y ) be defined at stage s1 with use &,(y) = s1. In the construction, it is possible that an M or 7-strategy below [ ^ ( n , m ) can enumerate a number x < s1 into B , at stage s2 > s1 say, and hence according to the construction, A [ , ( y ) is undefined at stage s2. If z remains in B , for any s > s2, B, 1 s1 # B,, 1 s1, and thus, the definition of A t , ( y ) [ s 1 ] cannot be recovered to that of stage s1. However, it is possible that at a stage s3 > s2, AE,(y) is redefined with use S<,,(y) = s3 and that later, at stage s4 say, z is moved out of B. Then because B r s l is recovered to B,, 1 s1, by the use principle, A[,(y)[s4] is recovered to the one at stage s1 with use &,,(y)[s4] = s1 of course. Also note that the definition of A f , ( y ) [ s 3 ] with use s3 is now undefined automatically because of the extraction of z. Thus, the M , 7-strategies below ( ^ ( n , m) do not make the definition of A?, inconsistent.
457 2.7. Q-strategies and substmtegies
Q,-strategy, then Uc
< say, attempts to construct a c.e. set Uc such that if We = OF,
ST B and for all i E w,Uc # a,”(. Define
k?(<, s) = m d x < s : (tJY < 4(@;(Y)[sl -1= Wc,s(y))}, r n ~ ( < , s=) max{O,lQ(<,t) : t < s & t is a <-stage}. Stage s is <-expansionary if s = 0 or s is a <-stage and &(<, s) > me(<,s). has two outcomes: 0
q,i U<# 07, (2.8) Let x be an q,i-strategy. Then x tries to figure out an inequality point between Uc and a,”( or between We and 0; (the latter is a global win). x proceeds as follows:
(1) Choose z(x)as a fresh number. (2) Wait for a stage s such that n:+(z(x)) $= 0, and W C ,1~ wi,s(z(x)) = 0:: i u i , s ( x ( x ) ) . (If this never happens, then x ( x ) is a witness t o the success of x , i . ) (3) Put z(x)into Uc and B. Protect B 1s from other strategies. W (4) Wait for a stage s‘ such that Oi,;+’ (x(x))$= 1. (If this never happens, then again x(x) is a witness t o the success of q , i . If it happens, then the change in (x(x)) between stages s and s’ can only be brought about by a change in We rwi,,(x(~)), which is irreversible since We is a c.e. set.)
fly
(5) Remove z(x)from B and protect B 1s from other strategies. (Now
is a permanent witness t o the success of x,i because OF rwi,s(x(x)) # We T~i,~(zC(x)). Note that taking x(x) from B leads t o a global win o n Q,. Uc is n o longer needed, and so we don’t need to care about the loss of B-permisssion f o r x(x) (which is left in Uc).) X(X)
1. Note that if x reaches step ( 5 ) , then an inequality between W, and O < ( B )is created, and so gets a global win for 5.
x has one outcome
x
458
3. Construction
Before describing the construction, we define the priority tree, T say, effectively. First define the priority of the requirements as follows:
6 < PE
<
Definition 3.2. (1) Define the root node X as a PE-strategy; (2) The immediate successors of a node are the possible outcomes of the corresponding strategy; (3) For 6 E T , 6 works for the highest priority requirement which is not satisfied at 6. For 6 E T , if 6 is one the PE-strategies, PD-strategies, M-strategies, and S-strategies, then z(6) will be defined. Furthermore, if 6 is a PEstrategy or an R-strategy, then k(6) will be defined. In the construction, if 6 is initialized, then any parameters will be cancelled. If 6 is reset, then any parameters, except k(S), will be cancelled. If 6 < 6' and 6 is reset or initialized, then 6' will be initialized simultaneously and automatically.
459
Construction
Without loss of generality, suppose that K is enumerated at 3 n + l stages, C is enumerated at 3 n + 2 stages. and that exactly one element is enumerated into K and C at each such a stage. Stage 0: Set B = D = E = 0, and initialize all nodes.
Stage s
+ 1:
(I) s _= 0 (mod 3). Let k be the number in K,+1 - K,. For any strategy 6, if k ( S ) > k, reset 6. If r(C,D; k ) [ s ]4,then enumerate y(k) into D. Otherwise, let z be the least y such that r(C,D; y) f. If r(C, D; z) is first defined, then define r ( C , D ;z)[s 11 = K,+l(z) with r(z)[s 11 fresh. Otherwise, let t be the last stage at which r(C,D; z)[t]4.If one of the following holds, then define r(C,D; z)[s 11 = K,+l(z) with y(x)[s 11 fresh.
+
+
+
+
(a) There is some y < k with y(y)[s] > y(k)[tJ; (b) There are some y-markers enumerated into D or removed from D since stage t. (c) C has a change below the largest active requesting number, z say, and z 5 y(k)[t].
If (a)-(c) do not apply then define r(C, D ;x)[s + 11 = 7(3)[tI* In any case, go t o the next stage.
+ 11 = K,+1 (x)with use
Y(Z"
(11) s
= 1 (mod 3).
Let c be the number in Cs+l- C,. If c is less than some active requesting number z = y ( k ( ~ ) ) [ swith ' ] s' < s, then declare that z's request is realized, and if y(k(~))[s]4,then let r(C, D ;k(~))[s 11 be undefined.
+
(111) s _= 2 (mod 3).
+
Say that a strategy 6 is visited at stage s 1, if 6 is eligible to act at a substage t of stage s + 1. First, let A, the root node, be eligible to act at substage 0.
+
Substage t: Let S be eligible to act at substage t. If t = s 1, then define cS+l= 6, initialize all C > gs+l,and go to the next stage. Otherwise, there are 7 cases: Case I. 6 = a is a PE-strategy.
al. If k(a)f, then define k(a)as a fresh number, define Ss+l = a , initialize all nodes with lower priority and go to the next stage.
460
a2. If k(a) J, z ( a ) t, then define .(a) as a fresh number, define 68+1= a , initialize all nodes with lower priority and go to the next stage. a3. If .(a) J and .(a) E E , then let a - ( d ) be eligible to act at the next substage; a4. If a is not satisfied, and a has a requesting number realized between the last a-stage and s + 1, i.e., there is some z with ga(z) # Cs+l(z).Let z be the least such disagreement point, and let s' 5 s be the stage at which ga(z) is defined. Move out all numbers enumerated into D after stage s' (including y ( k ( a ) ) [ s ' ] ) , put .(a) into E and s' into B , and for all y 2 k(a),let r(C,D;y) be undefined. For all $, and n with <^(n,m) a , if b&(m) > s' then let 6!,(m) be undefined. Declare that a is satisfied via z. Define os+l = a and initialize all nodes with lower priority. Go to the next stage. a5. Otherwise, suppose that a is in cycle (n). There are two subcases: Subcase 1. If @ f e D ( z ( a ) )+ [ s11 J= 0, then enumerate y ( k ( a ) ) [ s+ 11 into D, and for y 2 k(a),undefine r(C,D;y)and for z < y ( k ( a ) ) [ swith ] g t ( z ) t , define g t ( z ) = C,+l(z). Declare that y(k(a))[s] requests that C have a change below it to perform the Friedberg-Muchnik strategy. Stop cycle ( n ) and start cycle (n 1). Define o,+1 = and initialize all nodes with lower priority. Go to the next stage. Subcase 2. Otherwise, let a^(n) be eligible to act at the next substage;
+
Case 2. 6 =
is a PD-strategy.
z(P) t, then define z(P) as a fresh number, define &,+I = 0, initialize all nodes with lower priority and go to the next stage. 02. If z(P) J and z(P) E D , then let ,f3^(0) be eligible to act at the next substage; 03. If z(P) J, @;(z(P))[s 11J= D,(z(p)) = 0, then enumerate z(@) into D , define 6,+1 = fi, initialize all nodes with lower priority. Declare that ,B is satisfied and go to the next stage. p4. Otherwise. Let ,f?-(l)be eligible to act at the next substage.
01. If
+
Case 3. 6 = q is an M-strategy.
q l . If z(q) f, then define z(q) as fresh. Define 6, = q, initialize all strategies with lower priority and go to the next stage.
461
172. If ~ ( 1 7 )1,@ r Y ( x ( v ) -1= ) 0 = B*(x(v)), and w7,s t,4Bs;e(17),~(17),S) =
QCt O s ; e ( 1 7 ) , 4 1 7 ) , s ) , c
then enumerate x(q) into B , For all and n with <-(n,co) & a , if 6&(m) > x(q) then let 6&(m) be undefined. Define 6, = 17, initialize all strategies with lower priority and go to the next stage. 173. If x(q) 1, (~(17)) ./.= 1 = Bs(x(7/))and
@FFs
w7,st4Bs-;e(17),417),s-)=
Q?,
ru(BS-;.(17>,.(17),s-),
where s- is the stage at which x(q) in enumerated into B , then take x(q) out of B , define 6, = 17. (Note that by the use principle, those A&(m) undefined by the enumeration of x ( v ) are defined now because of the recovering of B 1 S[,(m).) Declare that 17 is satisfied, initialize all strategies with lower priority. Go to the next stage. 174. If 17 is satisfied, then let ~ ~ ( be0 eligible ) to act at the next substage. 175. Otherwise, let q n ( l ) be eligible to act at the next substage. Case 4.6 = o is an Af-strategy.
+
01. If s 1 is not 0-expansionary, then let un( f) be eligible to act at the next substage. 02. If s 1 is o-expansionary, and no substrategy requires attention, then let u-(d) be eligible to act at the next substage. 03. Otherwise, let i be the least number such that Su,i requires attention, and let Su,i receive attention as follows:
+
5 v, where v is the stage at which the gap is opened, define pg(m) = s 1. Let x be chosen by ( 0 2 ) at stage w. If C has a change below x between stages v and s 1,then declare that the gap is closed successfully and that Su,i is satisfied. Otherwise, declare that the gap is closed unsuccessfully. Define o ~ = + a-(gi), ~ initialize all nodes to the right of ~ , + l , and go to the next stage. Subcase 2. If 0 is inside a cogap and there is some y E dom($,,i) with C,,l %(Y)[V] # CIJ cpu(Y)[vl, where 'u is the last Qexpansionary stage, then let on(&) be eligible to act at the next substage. Subcase 1. If Su,i is inside a gap, then close the gap, and for all m
+
r
+
r
462
Subcase 3. Otherwise, Su,i is ready to open a gap. That is, there is some z entering Wi between stages w and s + 1, where w is the last an(gi)-stage, and there is some y < l ( g , s + l ) with $u,i(y)[s+ 11 t and z > max{u(Cs+l; e ( o ) , y ‘ , s 1) : y‘ 5 y}. Choose y as the least such number and define $u,i(y) = cPF(y)[s 11. Let on(gi) be eligible to act at the next substage.
+
Case 5. S =
+
< is an R-strategy.
s’ then let dF,,(rn) be undefined. Declare that 5 is satisfied via z. Define us+l = and initialize all nodes with lower priority. Go to the next stage. 54. Otherwise, suppose that is in cycle (n). There are three cases:
<
+
<
+
<
Subcase 1. s 1 is not <-expansionary. Then, let En(n, w)be eligible to act at the next substage. Subcase 2. s 1is <-expansionary and A f n is correct. Then, for any y < ~ R ( < , swith ) A ~ ~ ( Y ) [t,s ]define A ~ ~ ( Y + > [11s = ~ t , S + l ( y ) with use & ~ , ~ ( = y )s 1. Let En(n,cm) be eligible t o act at the next substage. Subcase 3. s 1 is J-expansionary and A t n is incorrect at (the least) y. Put y(k(())[s] into D. Declare that y(k([))[s] requests that C have a change below it to perform the disagreement strategy. For y 2 k(<), let r(C,D;y) be undefined and for z < r(k(t))[sI, if g t ( z ) t, define gt(z) = cs+l(z).Stop cycle (n)and start cycle ( n + 1). Define gS+l = and initialize all nodes with lower priority. Go to the next stage.
+
+
+
<
Case 6.S =
< is a Q-strategy.
463
+ 1 is <-expansionary, then let C-(O) be eligible to act at the next substage. C2. Otherwise, let <-(1) be eligible to act at the next substage.
Case 7. S = x is an z,i-strategy.
x1. If z(x) t, then define z(x) as a fresh number, define 6, = initialize all strategies with lower priority, go to the next stage. x2. I f 4 X ) 4,n,W.(z(X))[SI 4= 0 = UX,S(4X>) and wx,s t w x , s ( 4 x ) ) = s:;@
x,
t%MX)),
then enumerate z(x) into B and U,. For all E and n with <-(n,m) a,if S[,(m) > z(v) then let S:,(m) be undefined. Define 6, = x, initialize all strategies with lower priority and go to the next stage. ~ 3 If . z(x) -1, n ~ ( z ( x ) ) [ $= s ] 1 = Ux,,(z(x)),then take z(x) out of B. (Note that b y the use principle, those A$,(m) undefined by the enumeration of ~ ( 7 are ) defined again because of the recovering of B 1G;,(rn).) Define 6, = x, initialize all strategies with lower priority, go to the next stage. x4. Otherwise, let ~ ~ (be1eligible ) to act at the next substage. This completes the construction. 4. The Verification
In this section, we verify that the construction described above satisfies all the requirements. First we have:
Define f = liminf,
63,.
f is the true path of the construction.
464
Lemma 4.2. If C is noncomputable and has a cappable degree, then f o r
6 c f, ( I ) 6 is initialized or reset only finitely often; (2) 6 acts only finitely often. Proof: We prove the lemma by induction. Let 6- be the immediate
predecessor of 6. By the induction hypothesis, there is some (least) stage so such that (a) 6- cannot be initialized or reset afterwards; (b) 6- does not act afterwards. Since 6 is on f , there is some stage s1 2 SO such that for all s 2 sl(6, 2 6). By the choice of so, and the fact that 6 acts only when 6 is accessible, 6 cannot be initialized after s1. If 6 is a PE-strategy or an R-strategy, let k(6) be defined at stage s2 2 s1. Then k(6) cannot be canceled later and 6 can be reset only finitely often. (1) holds.
For (2), there are seven cases: (i) 6 = a is a PE-strategy. Suppose that (2) fails at a. Then a acts infinitely often. Thus, a can never be satisfied, and whenever a acts, a does the same thing: (1) puts the number y ( k ( a ) )into D,(2) extends the definition of ga, (3) starts the next cycle. Hence, ga is defined totally. We show below that ga computes C correctly. Thus, C is computable. A contradiction. Therefore, a can act only finitely often. (2) holds. Suppose not. Then there is a stage s3 > s2 such that some x with ga(x)[s3] .J.=0, and x enters C at this stage (ga becomes incorrect at x). Suppose that ga(x) is defined at stage s. Let s4 2 s3 be the least a-stage. Then by the construction, at stage s4, (Y takes y ( k ( a ) ) [ s ]out of D ,and puts s into B , %(a)into E. Thus,
E ( z ( a ) )= 1 # 0 = @ p ( x ( a ) ) [ S ] = G p D ( z ( a ) ) [ s q= ] @E@D(x(a)). a is satisfied, contradicting our assumption. (ii)
= P is a PD-strategy.
It is a standard Friedberg-Muchnik strategy. If there is no P-stage s > s1 such that @F(z(P))[s].J.=0, then P has no further action after stage s1. (2) holds. In this case, D(x(P)) = 0 # @F(x(p)), p is satisfied. Otherwise, let
465 s3 > s2 be the least ,&stage at which @;(x(p)) $= 0. Then p puts x ( p ) into D and at the same time initializes all strategies with lower priority. By the choice of s1, x(p) remains in D and the computation @;(x(p ))is preserved forever. Therefore,
p is satisfied, and ,6 has no more actions. (2) holds. (iii) 6 = 7 is an M-strategy. Let s3 > s1 be the stage at which z(7) is defined. Without loss of generality, suppose that at stage s4 > s3, 7 observes the following equations: @?(x(rl))[s4] -1= 0 = Bs,(x(7))and l~ , , ~ , ( 4 7 ) )= W,,,, t (P,+.,(x(q)). Then 17 enumerates z(q) into B and initializes all lower prior) ) &, l~,,~,(47)). Suppose ity strategies. Thus, - ( 4 7 ) ) t ~ , , ~ , ( ~ ( r l = that at an 7-stage s5 > s4, @F(x(q))[s5] $= 1 and iJ?& lCp,,,,(x(v)) = W,,,, r(p,,,s4 (~(17)). (If there is n o such stage, then a is satisfied by the first attempt, and 7 has n o more actions. (2) holds.) Then, W , 1 ( P , , ~ ~ ( Z ( ~ ) ) changes between stages s4 and s5, and hence @: l~p,,~,(x(q))must also have a change. The only reason causing such changes is the enumeration of z(7) into B , and x(7) is less than ~ q , , 4 ( ( p q , s 4 ( x Taking ( ~ ) ) . x(7) out of B at stage s 5 recovers the computations Q: l~p,,,~ ( ~ ( 7 )to) those computations at stage s 4 , and so creates a permanent inequality between W, and q:. r] is satisfied and does not act anymore. (2) holds.
q2,
(iv) 6 = CT is an N-strategy. CT
does nothing during the whole construction. (2) follows immediately.
(v) 6 = [ is an R-strategy.
<
Suppose that acts infinitely often. Then since each cycle of acts at most once (otherwise, this cycle will find a disagreement between @ f e D and We, satisfying E, and hence, E does not act anymore, (2) holds), we know that 5 starts infinitely many cycles, and each cycle extends the definition of gc properly. As a consequence, gc is defined totally. We now show that gt computes C correctly. Suppose not. Let y be the least disagreement between gc and C. Assume that g t ( y ) is defined at a stage s4 > s2. That is, there is a (least) z confirmed as an error at stage s 4 and -y(k(<))[q] is enumerated into D and requests that C have a change below it to perform the disagreement strategy. Let sz be the stage at which A?(Z) is defined. By gc(y) # C ( y ) ,-y(k(<))[s4]’srequest is realized later and by moving out all
466
numbers enumerated into D after stage s z , we get a disagreement between @ F e D ( z )and W t ( z )successfully. As a consequence, no more cycles can be started later. A contradiction. Thus, C is computable, contradicting the assumption that C has a cappable degree. (2) holds.
<
(vi) 6 = is a &-strategy. [ does nothing during the whole construction. (2) follows.
(vii) 6 = x is a 7-strategy.
<
Let C x be the Q-strategy for which x is working. Let z(x) be defined at stage s3 > s2. Without loss of generality, suppose that at a X-stage s4 > s3, we have
@p(z(x))[s4l-l= 0 Then x enumerates z(x) into B and i.7, at stage s4, and initializes all lower priority strategies. By the assumption that has infinitary outcome, we know @ p ( z ( x ) )cannot converge t o 1 at any larger X-stage, because otherwise, x will remove z(x) from B , making [-(1) f . Therefore, Uc(z(x)) = 1 # x is satisfied. After stage s4, x does not act 0 anymore. (2) holds.
<
@F(z(x)),
c
Lemma 4.3. If C i s noncomputable and has a cappable degree, then for f. Therefore, I f I = co. any 6 C_ f , there is some 0 such that 6 - 0
Proof: Let 6 C f be any strategy on f . We only need to consider the cases when 6 is a PE-strategy, an N-strategy or an R-strategy. Case I: 6 = a is a PE-strategy. By Lemma 4.2, if C is not computable, then a acts a t most finitely often. Let s be the last stage at which Q acts and suppose that at stage s, a is running cycle (n). There are two subcases: Subcase 1: .(a) is enumerated into E at stage s. By the construction, we know that some y with g a ( y ) $= 0 enters C between SO, the stage at which g a ( y ) is defined, and stage s. Now a puts .(a) into E , and simultaneously puts SO into B , taking r ( k ( a ) ) [ s oout ] of D.The enumeration of .(a) makes E(z(a))= 1, and the removal of y(k(a))[so] recovers the computation @EeD(z(a)) to @ E @ D ( z ( ~ ) )(Note [ s ~ ] that . the enumeration of so into B does n o t change the computation @E@D(z(a))[so] because (p,(z(a))[so]< SO.) By preserving this computation, we have
E(a:(a))= 1 # 0 = @ E e D ( ~ ( ~ ) ) [=~ @ o ] E@D(z(~)).
467
a is satisfied, and any a-stage after stage s is also an a-(d)-stage. a - ( d ) g
f.
Subcase 2: a starts cycle (n + 1). In this case, a puts ~ ( k ( a ) ) [into s ] D. By the choice of s, the compu) ) not converge to 0 at any larger a-stage, because tation @ z @ D ( x ( adoes otherwise, a will act again. Thus, any a-stage after stage s is also an a-(n + 1)-stage, a-(n + 1) C: f . Case 6 = a is an N-strategy. If there are only finitely many a-expansionary stages, then after a stage large enough, every a-stage is an a ^ ( f)-stage, and hence a-(f) C f . If there are infinitely many a-expansionary stages, and after a stage large enough, a has no substrategy requiring attention anymore, then every a-expansionary stage is a u-(d)-stage. a n ( d ) f . Otherwise, there are infinitely many a-expansionary stages, and a's substrategies can require attention infinitely often. We claim that there is some i such that a-(gi) C f or a ^ ( & ) C f . Suppose not. Then for each i, ah(gi) and an(&) can be visited only finitely often. We show below that each S,,i-requirement is satisfied, and hence C has a promptly simple degree. Contradicting our assumption that C has a cappable degree. Fix i. Then we can assume that after stage SO large enough, no substrategy with higher priority can require attention anymore. Then, after stage S O , whenever S,,i requires attention, Su,i can receive attention. Without loss of generality, assume that Wi is infinite. By our assumption that a-(gi) is not on f , So,i can open gaps finitely often. Thus we can assume that after stage s1 2 SO, So,i opens no gap. Then, dom(+,,i) = dom($~,,i[s~]) is finite. Moreover, for each x E dorn($Jg,i),a:(.) converges (otherwise, Sm,i will require attention infinitely often, making a^(&) g f ) . Let t = max{cp,(y) : y E dorn(&,i)} and let s2 2 s1 be the stage with C, 1z = C z. We claim that at stage s2, S,,i is satisfied or inside a gap, because otherwise, since Wi is infinite, there will be an a-expansionary stage s3 > s2 at which S,,i requires attention to open a gap, which is impossible by the choice of s1. In the first case, we are done. In the second case, since the gap cannot be canceled, the gap will be closed at the next a-expansionary stage s4 > s2, and furthermore, So,iis satisfied by stage s4. Case 3: 6 = 6 is an R-strategy. The argument in the proof of Lemma 4.2 shows that 5 can start only finitely many cycles. Let cycle ( n ) be the largest one being started in the
c
r
468
whole construction. Suppose that cycle (n) is started at stage s1. Then, after stage s1, for m < n, no ["(m, -) can be visited again. There are two subcases:
Subcase 1: There are only finitely many J-expansionary stages. If there is some (least) y such that g&) J. and gt(y) # C(y), where gt(y) is defined by cycle (m) 5 ( n ) at stage s2, then by the construction, we will take y(k(())[s2], together with other numbers, out of D to execute the disagreement strategy. E is satisfied and hence, any J-stage larger than s2 is a <"(d)-stage. (-(d) g f . Otherwise, let s3 > s1 be the last J-expansionary stage. Then any J-stage s > s3 is also a E-(n, w)-stage. <"(n,w) g f . Subcase 2: There are infinitely many &expansionary stages. Then A:n will be defined infinitely often, and for any y, if A:,(y)[s] is defined, then A f , ( y ) [ s ] = lV~,~(y),because otherwise, cycle ( n )would start cycle (n+l),contradicting our choice of n. Thus, each <-expansionary stage larger than s1 is a ["(n, co)-stage, (-(n, co) g f . 0 Lemma 4.4. I f C is noncomputable and has a cappable degree, t h e n for any 6 c f ,
(1) I f (2) I f (3) I f (4) I f (5) I f (6) I f
6 is a P-strategy, t h e n P6 i s satisfied. 6 = q i s a n M - s t r a t e g y , t h e n M , i s satisfied. 6 = u is a n N-strategy, t h e n Nu is satisfied. 6 = E is a n R-strategy, t h e n R.6 is satisfied. 6 = is a Q-strategy, t h e n Qc is satisfied. 6 = x is a 7-strategy, t h e n Tx is satisfied.
Proof: (1)By the proof of (i), (ii) in Lemma 4.2 and (i) in Lemma 4.3. (2) By the proof of (iii) in Lemma 4.2. (3) Suppose that = = h, is total. Then there are infinitely many a-expansionary stages, and by Lemma 4.3, there is some i such that un(gi) C f or U-(di) C f . We now prove that a"(di) cannot be on f . Otherwise, dom(+,,i) is finite. Let s1 is the last stage at which &,i opens a gap. Thus, for any u-(di)-stage s > s1,there is some y E dom(l(lo,i)such that the computation @:(y) changes. Since dom($,,i) is finite, there is some (least) y E dom($,,i) such that $F(y)t, and hence h, is not total. A contradiction. Thus, uh(gi) c f , and hence, as described in the N-strategy, 1Clu,i is totally defined, and computes h, correctly. u is satisfied.
469
@FeD
(4) Suppose that = We. Then there are infinitely many [expansionary stages. By Lemma 4.3, we know that there is some (least) i such that ["(i,00) is on f . We now show that if is total, then the p.c. functional A; is totally defined and computes We correctly. First note that if A? is defined, then A? = We, because otherwise, [ will start to perform the disagreement strategy. There are two possibilities. One is that [ starts a new cycle, and the other one is that 0 performs the disagreement strategy at last. Both make [-(i, 00) not accessible anymore. A contradiction. We now show that A t is totally defined. Fix x. Let A f i ( x ) be defined at stage so. Then S,,i(z)[so]= SO. W.l.o.g., suppose that at stage s1 > SO, some strategy T > e-(i,00) enumerates a number z < SO into B and undefines A E i ( x ) . Obviously, T is below ("(i, co). Then at the next [expansionary stage s2, A t i ( x ) is redefined with use Sc,i(z)[s2]= s2. By the construction, all strategies with priority lower than T are initialized at stage s1 and cannot redefine A,'li(x)[s2] again. Since only finitely many nodes between <"(i,co) and [' have been visited before stage s2, and only these strategies' actions can undefine the newly defined A C i ( x ) ,there is a large enough stage s, at which A f i ( x )is defined and this definition cannot be injured afterwards. (Note t h a t if T is a n M or a 7 - s t r a t e g y , t h e n z can be removed f r o m B f o r t h e sake of t h e r-strategy, and if so, t h e n t h e definition of A t i ( x ) will be recovered t o A f i ( z ) [ s o ]by t h e u s e principle. Again, since there are only finitely m a n y strategies between e"(i, 00) and T, such a procedure can happen finiteZy often.) Then $= A,'li(z)[s,]. AEi(z)is defined. (5) Suppose that Wc = O f , then there are infinitely many Iexpansionary stages and so I-(O) C f . Uc is constructed by C's substrategies 'Fi, i € w,which are arranged below C"(0) by the construction of T . First, we show that Uc ST B. In the construction, when z(x) is enumerated into Uc, by x at stage SO say, z ( x ) is also enumerated into B. Note that x(x) can only be taken out by x , at stage s2 say, and such an extraction action creates an inequality
Remember that x's action initializes all strategies with lower priority, particularly, those _> Sn(l). Let s3 > s2 be any I-stage, then by the choice of s2, s3 is not <-expansionary, and so Cn(l) is visited at stage s3. This is why x cannot be initialized later, and hence the computation O f ( x ( x ) ) [ s 2=] 0
e
470
is preserved forever. <-(1) c f . A contradiction. ( 6 ) By the proof of (vii) in Lemma 4.2.
0
Lemma 4.5. F(C,D) = K .
Proof: By Lemma 4.1, r(C,D ) is a p.c. functional. By actions during 3n 1 stages, if r(C,D) is total, then r(C, D ) = K . Fix k . Let 6 = f [ k . Then 6 can only be visited after stage k . Since 6 is on the true path, by Lemma 4.2, there is some (least) stage sg such that 6 cannot be initialized or reset afterwards. Let s1 2 sg be the stage at which k ( 6 ) is defined. Then k ( 6 ) > k and k ( 6 ) cannot be canceled later. Let sz 2 s1 be the least stage such that K 1( k ( 6 ) 1) = Ks,t ( k ( 6 ) 1). Let s3 > sz be the stage at which r(C, D ; k ( 6 ) ) is defined. Then for any 1 < k ( 6 ) , y(Z)[s3] < y(k(6))[s3],and D has no change below y(k(6))[s3] afterwards.
+
+
+
D Ty(k(6))[~3I= Dss ~ Y ( ~ ( S ) ) [ S Q I . Furthermore, since C is noncomputable, there are only finitely many stages wo,v1,... ,w, such that y(k(d))[vi],i 5 n , can request C have a change below them (i.e., y ( k ( b ) ) [ v i ]are enumerated into D to lift y(k(6))). Let s4 2 v, be the least stage after which y(k(6)) does not request anymore. Then either (T is satisfied at stage s4 (one of the y(k(6))’srequests is realized in this case) or C will have no change below y(k(G))[vi] for any i 5 n. In both cases, if r(C, D; k ( 6 ) ) is defined at stage s5 2 s4, then for all s 2 s5, y(k(S))[s] = y(k(d))[s4] by the G-strategy. Therefore, r(C,D; k ( 6 ) ) 4,and hence r(C, D ; k ) 4. 0 Acknowledgement: This project is supported by New Zealand FRST Post-Doctoral Fellowship. References 1. K. Ambos-Spies, C. G. Jockusch, Jr., R. A. Shore and R. I. Soare, An algebraic decomposition of the recursively enumerable degrees and the coincidence of several degree classes with the promptly simple degrees, Trans. Amer. Math. SOC.281 (1984), 109-128. 2. M. M. Arslanov, Structural properties of the degrees below 0’, Dokl. Akad, Nauk SSSR(N. S.) 283 (1985), 270-273. 3. M. M. Arslanov, S. Lempp and R. A. Shore, O n isolating r.e. and isolated d.r. e. degrees, in “Computability, Enumerability, Unsolvability” (Cooper, Slaman, Wainer, eds), 1996, 61-80.
471 4. S. B. Cooper, Degrees of Unsolvability, Ph. D. thesis, Leicester University, Leicester, England. 5. S. B. Cooper, L. Harrington, A. H. Lachlan, S. Lempp and R. I. Soare, T h e d.r.e. degrees are n o t dense, Ann. Pure Appl. Logic 55 (1991), 125-151. 6. S. B. Cooper, S. Lempp and P. Watson, Weak density and cupping in t h e d-r.e. degrees, Israel J. Math. 67 (1989), 137 -152. 7. S. B. Cooper and X. Yi, Isolated d.r.e. degrees, University of Leeds, Dept. of Pure Math., 1995, Preprint series, No. 17, 25pp. 8. D. Ding and L. Qian, Isolated d.r.e. degrees are dense in r.e. degree structure, Arch. Math. Logic 36 (1996), 1-10, 9. D. Ding and L. Qian, Lattice embedding into d-r.e. degrees preserving 0 and 1, in “Proceedings of the Sixth Asian Logic Conference” (C. T. Chong, Q. Feng, D. Ding, Q. Huang and M. Yasugi, eds) (1998), 67-81. 10. R. G. Downey, D.r.e. degrees and the nondiamond theorem, Bull. London Math. SOC.21 (1989), 43-50. 11. R. G. Downey, A. Li and G. Wu, Every cappable c.e. degree i s complemented in the d.c.e. degrees, in preparation. 12. S. Ishmukhametov, D.r.e. sets, their degrees and index sets, Thesis, Novosibirsk, Russia, 1986. 13. S. Ishmukhametov and G. Wu, Isolation and the high/low hierarchy, Arch. Math. Logic 41 (2002), 259-266. 14. A. H. Lachlan, Lower bounds for pairs of recursively enumerable degrees, Proc. London math. SOC.16 (1966), 537-569. 15. G. LaForte, T h e isolated d.r.e. degrees are dense in the r.e. degrees, Math. Logic Quart. 42 (1996), 83-103. 16. A. Li and X. Yi, Cupping the recursively enumerable degrees by d.r.e. degrees, Proc. London Math. SOC.78 (1999), 1-21. 17. R. I. Soare, Recursively Enumerable Set s and Degrees, SpringerVerlag, Berlin, 1987. 18. G. Wu, Nonisolated degrees and the j u m p operator, Ann. Pure Appl. Logic, 117 (2002), 211-223. 19. G. Wu, Isolation and diamond embeddings, J. Symbolic Logic 67 (2002), 1055-1064. 20. G. Wu, O n the density of the pseudo-isolated degrees, Victoria University of Wellington, School of Mathematical and Computing Sciences, 2002, Research Report, No. 02-10, 16pp. 21. G. Wu, Structural properties of the d.c.e. degrees and presentations of c.e. reals, Ph.D. Thesis, Victoria University of Wellington, 2002. 22. G. Wu, Embedding s8 into the d.c.e. degrees preserving 0 and 1, in preparation.