Library Cpdt.GeneralRec
Well-Founded Recursion
We have a set equipped with some "less-than-or-equal-to" test.
A standard function inserts an element into a sorted list, preserving sortedness.
Fixpoint insert (x : A) (ls : list A) : list A :=
match ls with
| nil => x :: nil
| h :: ls' =>
if le x h
then x :: ls
else h :: insert x ls'
end.
We will also need a function to merge two sorted lists. (We use a less efficient implementation than usual, because the more efficient implementation already forces us to think about well-founded recursion, while here we are only interested in setting up the example of merge sort.)
Fixpoint merge (ls1 ls2 : list A) : list A :=
match ls1 with
| nil => ls2
| h :: ls' => insert h (merge ls' ls2)
end.
The last helper function for classic merge sort is the one that follows, to split a list arbitrarily into two pieces of approximately equal length.
Fixpoint split (ls : list A) : list A * list A :=
match ls with
| nil => (nil, nil)
| h :: nil => (h :: nil, nil)
| h1 :: h2 :: ls' =>
let (ls1, ls2) := split ls' in
(h1 :: ls1, h2 :: ls2)
end.
Now, let us try to write the final sorting function, using a natural number "<=" test leb from the standard library.
Fixpoint mergeSort (ls : list A) : list A :=
if leb (length ls) 1
then ls
else let lss := split ls in
merge (mergeSort (fst lss)) (mergeSort (snd lss)).
The definition is rejected for not following the simple primitive recursion criterion. In particular, it is not apparent that recursive calls to mergeSort are syntactic subterms of the original argument ls; indeed, they are not, yet we know this is a well-founded recursive definition.
To produce an acceptable definition, we need to choose a well-founded relation and prove that mergeSort respects it. A good starting point is an examination of how well-foundedness is formalized in the Coq standard library.
Fixpoint mergeSort (ls : list A) : list A :=
if leb (length ls) 1
then ls
else let lss := split ls in
merge (mergeSort (fst lss)) (mergeSort (snd lss)).
Recursive call to mergeSort has principal argument equal to "fst (split ls)" instead of a subterm of "ls".
Print well_founded.
well_founded =
fun (A : Type) (R : A -> A -> Prop) => forall a : A, Acc R a
Print Acc.
Inductive Acc (A : Type) (R : A -> A -> Prop) (x : A) : Prop :=
Acc_intro : (forall y : A, R y x -> Acc R y) -> Acc R x
CoInductive infiniteDecreasingChain A (R : A -> A -> Prop) : stream A -> Prop :=
| ChainCons : forall x y s, infiniteDecreasingChain R (Cons y s)
-> R y x
-> infiniteDecreasingChain R (Cons x (Cons y s)).
We can now prove that any accessible element cannot be the beginning of any infinite decreasing chain.
Lemma noBadChains' : forall A (R : A -> A -> Prop) x, Acc R x
-> forall s, ~infiniteDecreasingChain R (Cons x s).
induction 1; crush;
match goal with
| [ H : infiniteDecreasingChain _ _ |- _ ] => inversion H; eauto
end.
Qed.
From here, the absence of infinite decreasing chains in well-founded sets is immediate.
Theorem noBadChains : forall A (R : A -> A -> Prop), well_founded R
-> forall s, ~infiniteDecreasingChain R s.
destruct s; apply noBadChains'; auto.
Qed.
Absence of infinite decreasing chains implies absence of infinitely nested recursive calls, for any recursive definition that respects the well-founded relation. The Fix combinator from the standard library formalizes that intuition:
Check Fix.
Fix
: forall (A : Type) (R : A -> A -> Prop),
well_founded R ->
forall P : A -> Type,
(forall x : A, (forall y : A, R y x -> P y) -> P x) ->
forall x : A, P x
forall x : A, (forall y : A, R y x -> P y) -> P x
We must prove that the relation is truly well-founded. To save some space in the rest of this chapter, we skip right to nice, automated proof scripts, though we postpone introducing the principles behind such scripts to Part III of the book. Curious readers may still replace semicolons with periods and newlines to step through these scripts interactively.
Hint Constructors Acc.
Lemma lengthOrder_wf' : forall len, forall ls, length ls <= len -> Acc lengthOrder ls.
unfold lengthOrder; induction len; crush.
Defined.
Theorem lengthOrder_wf : well_founded lengthOrder.
red; intro; eapply lengthOrder_wf'; eauto.
Defined.
Notice that we end these proofs with Defined, not Qed. Recall that Defined marks the theorems as transparent, so that the details of their proofs may be used during program execution. Why could such details possibly matter for computation? It turns out that Fix satisfies the primitive recursion restriction by declaring itself as recursive in the structure of Acc proofs. This is possible because Acc proofs follow a predictable inductive structure. We must do work, as in the last theorem's proof, to establish that all elements of a type belong to Acc, but the automatic unwinding of those proofs during recursion is straightforward. If we ended the proof with Qed, the proof details would be hidden from computation, in which case the unwinding process would get stuck.
To justify our two recursive mergeSort calls, we will also need to prove that split respects the lengthOrder relation. These proofs, too, must be kept transparent, to avoid stuckness of Fix evaluation. We use the syntax @foo to reference identifier foo with its implicit argument behavior turned off. (The proof details below use Ltac features not introduced yet, and they are safe to skip for now.)
Lemma split_wf : forall len ls, 2 <= length ls <= len
-> let (ls1, ls2) := split ls in
lengthOrder ls1 ls /\ lengthOrder ls2 ls.
unfold lengthOrder; induction len; crush; do 2 (destruct ls; crush);
destruct (le_lt_dec 2 (length ls));
repeat (match goal with
| [ _ : length ?E < 2 |- _ ] => destruct E
| [ _ : S (length ?E) < 2 |- _ ] => destruct E
| [ IH : _ |- context[split ?L] ] =>
specialize (IH L); destruct (split L); destruct IH
end; crush).
Defined.
Ltac split_wf := intros ls ?; intros; generalize (@split_wf (length ls) ls);
destruct (split ls); destruct 1; crush.
Lemma split_wf1 : forall ls, 2 <= length ls
-> lengthOrder (fst (split ls)) ls.
split_wf.
Defined.
Lemma split_wf2 : forall ls, 2 <= length ls
-> lengthOrder (snd (split ls)) ls.
split_wf.
Defined.
Hint Resolve split_wf1 split_wf2.
To write the function definition itself, we use the refine tactic as a convenient way to write a program that needs to manipulate proofs, without writing out those proofs manually. We also use a replacement le_lt_dec for leb that has a more interesting dependent type. (Note that we would not be able to complete the definition without this change, since refine will generate subgoals for the if branches based only on the type of the test expression, not its value.)
Definition mergeSort : list A -> list A.
refine (Fix lengthOrder_wf (fun _ => list A)
(fun (ls : list A)
(mergeSort : forall ls' : list A, lengthOrder ls' ls -> list A) =>
if le_lt_dec 2 (length ls)
then let lss := split ls in
merge (mergeSort (fst lss) _) (mergeSort (snd lss) _)
else ls)); subst lss; eauto.
Defined.
End mergeSort.
The important thing is that it is now easy to evaluate calls to mergeSort.
= 1 :: 2 :: 8 :: 19 :: 36 :: nil
Since the subject of this chapter is merely how to define functions with unusual recursion structure, we will not prove any further correctness theorems about mergeSort. Instead, we stop at proving that mergeSort has the expected computational behavior, for all inputs, not merely the one we just tested.
Theorem mergeSort_eq : forall A (le : A -> A -> bool) ls,
mergeSort le ls = if le_lt_dec 2 (length ls)
then let lss := split ls in
merge le (mergeSort le (fst lss)) (mergeSort le (snd lss))
else ls.
intros; apply (Fix_eq (@lengthOrder_wf A) (fun _ => list A)); intros.
The library theorem Fix_eq imposes one more strange subgoal upon us. We must prove that the function body is unable to distinguish between "self" arguments that map equal inputs to equal outputs. One might think this should be true of any Gallina code, but in fact this general function extensionality property is neither provable nor disprovable within Coq. The type of Fix_eq makes clear what we must show manually:
Check Fix_eq.
Fix_eq
: forall (A : Type) (R : A -> A -> Prop) (Rwf : well_founded R)
(P : A -> Type)
(F : forall x : A, (forall y : A, R y x -> P y) -> P x),
(forall (x : A) (f g : forall y : A, R y x -> P y),
(forall (y : A) (p : R y x), f y p = g y p) -> F x f = F x g) ->
forall x : A,
Fix Rwf P F x = F x (fun (y : A) (_ : R y x) => Fix Rwf P F y)
match goal with
| [ |- context[match ?E with left _ => _ | right _ => _ end] ] => destruct E
end; simpl; f_equal; auto.
Qed.
As a final test of our definition's suitability, we can extract to OCaml.
let rec mergeSort le x = match le_lt_dec (S (S O)) (length x) with | Left -> let lss = split x in merge le (mergeSort le (fst lss)) (mergeSort le (snd lss)) | Right -> x
Check well_founded_induction.
well_founded_induction
: forall (A : Type) (R : A -> A -> Prop),
well_founded R ->
forall P : A -> Set,
(forall x : A, (forall y : A, R y x -> P y) -> P x) ->
forall a : A, P a
A Non-Termination Monad Inspired by Domain Theory
The type A describes the result a computation will yield, if it terminates.
We give a rich dependent type to computations themselves:
Definition computation :=
{f : nat -> option A
| forall (n : nat) (v : A),
f n = Some v
-> forall (n' : nat), n' >= n
-> f n' = Some v}.
A computation is fundamentally a function f from an approximation level n to an optional result. Intuitively, higher n values enable termination in more cases than lower values. A call to f may return None to indicate that n was not high enough to run the computation to completion; higher n values may yield Some. Further, the proof obligation within the subset type asserts that f is monotone in an appropriate sense: when some n is sufficient to produce termination, so are all higher n values, and they all yield the same program result v.
It is easy to define a relation characterizing when a computation runs to a particular result at a particular approximation level.
On top of runTo, we also define run, which is the most abstract notion of when a computation runs to a value.
The book source code contains at this point some tactics, lemma proofs, and hint commands, to be used in proving facts about computations. Since their details are orthogonal to the message of this chapter, I have omitted them in the rendered version.
Now, as a simple first example of a computation, we can define Bottom, which corresponds to an infinite loop. For any approximation level, it fails to terminate (returns None). Note the use of abstract to create a new opaque lemma for the proof found by the run tactic. In contrast to the previous section, opaque proofs are fine here, since the proof components of computations do not influence evaluation behavior. It is generally preferable to make proofs opaque when possible, as this enforces a kind of modularity in the code to follow, preventing it from depending on any details of the proof.
Section Bottom.
Variable A : Type.
Definition Bottom : computation A.
exists (fun _ : nat => @None A); abstract run.
Defined.
Theorem run_Bottom : forall v, ~run Bottom v.
run.
Qed.
End Bottom.
A slightly more complicated example is Return, which gives the same terminating answer at every approximation level.
Section Return.
Variable A : Type.
Variable v : A.
Definition Return : computation A.
intros; exists (fun _ : nat => Some v); abstract run.
Defined.
Theorem run_Return : run Return v.
run.
Qed.
End Return.
The name Return was meant to be suggestive of the standard operations of monads. The other standard operation is Bind, which lets us run one computation and, if it terminates, pass its result off to another computation. We implement bind using the notation let (x, y) := e1 in e2, for pulling apart the value e1 which may be thought of as a pair. The second component of a computation is a proof, which we do not need to mention directly in the definition of Bind.
Section Bind.
Variables A B : Type.
Variable m1 : computation A.
Variable m2 : A -> computation B.
Definition Bind : computation B.
exists (fun n =>
let (f1, _) := m1 in
match f1 n with
| None => None
| Some v =>
let (f2, _) := m2 v in
f2 n
end); abstract run.
Defined.
Theorem run_Bind : forall (v1 : A) (v2 : B),
run m1 v1
-> run (m2 v1) v2
-> run Bind v2.
run; match goal with
| [ x : nat, y : nat |- _ ] => exists (max x y)
end; run.
Qed.
End Bind.
A simple notation lets us write Bind calls the way they appear in Haskell.
We can verify that we have indeed defined a monad, by proving the standard monad laws. Part of the exercise is choosing an appropriate notion of equality between computations. We use "equality at all approximation levels."
Definition meq A (m1 m2 : computation A) := forall n, proj1_sig m1 n = proj1_sig m2 n.
Theorem left_identity : forall A B (a : A) (f : A -> computation B),
meq (Bind (Return a) f) (f a).
run.
Qed.
Theorem right_identity : forall A (m : computation A),
meq (Bind m (@Return _)) m.
run.
Qed.
Theorem associativity : forall A B C (m : computation A)
(f : A -> computation B) (g : B -> computation C),
meq (Bind (Bind m f) g) (Bind m (fun x => Bind (f x) g)).
run.
Qed.
Now we come to the piece most directly inspired by domain theory. We want to support general recursive function definitions, but domain theory tells us that not all definitions are reasonable; some fail to be continuous and thus represent unrealizable computations. To formalize an analogous notion of continuity for our non-termination monad, we write down the approximation relation on computation results that we have had in mind all along.
Section lattice.
Variable A : Type.
Definition leq (x y : option A) :=
forall v, x = Some v -> y = Some v.
End lattice.
We now have the tools we need to define a new Fix combinator that, unlike the one we saw in the prior section, does not require a termination proof, and in fact admits recursive definition of functions that fail to terminate on some or all inputs.
Section Fix.
First, we have the function domain and range types.
Next comes the function body, which is written as though it can be parameterized over itself, for recursive calls.
Finally, we impose an obligation to prove that the body f is continuous. That is, when f terminates according to one recursive version of itself, it also terminates with the same result at the same approximation level when passed a recursive version that refines the original, according to leq.
Hypothesis f_continuous : forall n v v1 x,
runTo (f v1 x) n v
-> forall (v2 : A -> computation B),
(forall x, leq (proj1_sig (v1 x) n) (proj1_sig (v2 x) n))
-> runTo (f v2 x) n v.
The computational part of the Fix combinator is easy to define. At approximation level 0, we diverge; at higher levels, we run the body with a functional argument drawn from the next lower level.
Fixpoint Fix' (n : nat) (x : A) : computation B :=
match n with
| O => Bottom _
| S n' => f (Fix' n') x
end.
Now it is straightforward to package Fix' as a computation combinator Fix.
Hint Extern 1 (_ >= _) => omega.
Hint Unfold leq.
Lemma Fix'_ok : forall steps n x v, proj1_sig (Fix' n x) steps = Some v
-> forall n', n' >= n
-> proj1_sig (Fix' n' x) steps = Some v.
unfold runTo in *; induction n; crush;
match goal with
| [ H : _ >= _ |- _ ] => inversion H; crush; eauto
end.
Qed.
Hint Resolve Fix'_ok.
Hint Extern 1 (proj1_sig _ _ = _) => simpl;
match goal with
| [ |- proj1_sig ?E _ = _ ] => eapply (proj2_sig E)
end.
Definition Fix : A -> computation B.
intro x; exists (fun n => proj1_sig (Fix' n x) n); abstract run.
Defined.
Finally, we can prove that Fix obeys the expected computation rule.
Theorem run_Fix : forall x v,
run (f Fix x) v
-> run (Fix x) v.
run; match goal with
| [ n : nat |- _ ] => exists (S n); eauto
end.
Qed.
End Fix.
After all that work, it is now fairly painless to define a version of mergeSort that requires no proof of termination. We appeal to a program-specific tactic whose definition is hidden here but present in the book source.
Definition mergeSort' : forall A, (A -> A -> bool) -> list A -> computation (list A).
refine (fun A le => Fix
(fun (mergeSort : list A -> computation (list A))
(ls : list A) =>
if le_lt_dec 2 (length ls)
then let lss := split ls in
ls1 <- mergeSort (fst lss);
ls2 <- mergeSort (snd lss);
Return (merge le ls1 ls2)
else Return ls) _); abstract mergeSort'.
Defined.
Furthermore, "running" mergeSort' on concrete inputs is as easy as choosing a sufficiently high approximation level and letting Coq's computation rules do the rest. Contrast this with the proof work that goes into deriving an evaluation fact for a deeply embedded language, with one explicit proof rule application per execution step.
Lemma test_mergeSort' : run (mergeSort' leb (1 :: 2 :: 36 :: 8 :: 19 :: nil))
(1 :: 2 :: 8 :: 19 :: 36 :: nil).
exists 4; reflexivity.
Qed.
There is another benefit of our new Fix compared with the one we used in the previous section: we can now write recursive functions that sometimes fail to terminate, without losing easy reasoning principles for the terminating cases. Consider this simple example, which appeals to another tactic whose definition we elide here.
Definition looper : bool -> computation unit.
refine (Fix (fun looper (b : bool) =>
if b then Return tt else looper b) _); abstract looper.
Defined.
Lemma test_looper : run (looper true) tt.
exists 1; reflexivity.
Qed.
As before, proving outputs for specific inputs is as easy as demonstrating a high enough approximation level.
There are other theorems that are important to prove about combinators like Return, Bind, and Fix. In general, for a computation c, we sometimes have a hypothesis proving run c v for some v, and we want to perform inversion to deduce what v must be. Each combinator should ideally have a theorem of that kind, for c built directly from that combinator. We have omitted such theorems here, but they are not hard to prove. In general, the domain theory-inspired approach avoids the type-theoretic "gotchas" that tend to show up in approaches that try to mix normal Coq computation with explicit syntax types. The next section of this chapter demonstrates two alternate approaches of that sort. In the final section of the chapter, we review the pros and cons of the different choices, coming to the conclusion that none of them is obviously better than any one of the others for all situations.
There are two key downsides to both of the previous approaches: both require unusual syntax based on explicit calls to fixpoint combinators, and both generate immediate proof obligations about the bodies of recursive definitions. In Chapter 5, we have already seen how co-inductive types support recursive definitions that exhibit certain well-behaved varieties of non-termination. It turns out that we can leverage that co-induction support for encoding of general recursive definitions, by adding layers of co-inductive syntax. In effect, we mix elements of shallow and deep embeddings.
Our first example of this kind, proposed by Capretta, defines a silly-looking type of thunks; that is, computations that may be forced to yield results, if they terminate.
Co-Inductive Non-Termination Monads
A computation is either an immediate Answer or another computation wrapped inside Think. Since thunk is co-inductive, every thunk type is inhabited by an infinite nesting of Thinks, standing for non-termination. Terminating results are Answer wrapped inside some finite number of Thinks.
Why bother to write such a strange definition? The definition of thunk is motivated by the ability it gives us to define a "bind" operation, similar to the one we defined in the previous section.
CoFixpoint TBind A B (m1 : thunk A) (m2 : A -> thunk B) : thunk B :=
match m1 with
| Answer x => m2 x
| Think m1' => Think (TBind m1' m2)
end.
Note that the definition would violate the co-recursion guardedness restriction if we left out the seemingly superfluous Think on the righthand side of the second match branch.
We can prove that Answer and TBind form a monad for thunk. The proof is omitted here but present in the book source. As usual for this sort of proof, a key element is choosing an appropriate notion of equality for thunks.
In the proofs to follow, we will need a function similar to one we saw in Chapter 5, to pull apart and reassemble a thunk in a way that provokes reduction of co-recursive calls.
Definition frob A (m : thunk A) : thunk A :=
match m with
| Answer x => Answer x
| Think m' => Think m'
end.
Theorem frob_eq : forall A (m : thunk A), frob m = m.
destruct m; reflexivity.
Qed.
As a simple example, here is how we might define a tail-recursive factorial function.
CoFixpoint fact (n acc : nat) : thunk nat :=
match n with
| O => Answer acc
| S n' => Think (fact n' (S n' * acc))
end.
To test our definition, we need an evaluation relation that characterizes results of evaluating thunks.
Inductive eval A : thunk A -> A -> Prop :=
| EvalAnswer : forall x, eval (Answer x) x
| EvalThink : forall m x, eval m x -> eval (Think m) x.
Hint Rewrite frob_eq.
Lemma eval_frob : forall A (c : thunk A) x,
eval (frob c) x
-> eval c x.
crush.
Qed.
Theorem eval_fact : eval (fact 5 1) 120.
repeat (apply eval_frob; simpl; constructor).
Qed.
We need to apply constructors of eval explicitly, but the process is easy to automate completely for concrete input programs.
Now consider another very similar definition, this time of a Fibonacci number function.
CoFixpoint fib (n : nat) : thunk nat :=
match n with
| 0 => Answer 1
| 1 => Answer 1
| _ => n1 <- fib (pred n);
n2 <- fib (pred (pred n));
Answer (n1 + n2)
end.
CoInductive comp (A : Type) : Type :=
| Ret : A -> comp A
| Bnd : forall B, comp B -> (B -> comp A) -> comp A.
This example shows off Coq's support for recursively non-uniform parameters, as in the case of the parameter A declared above, where each constructor's type ends in comp A, but there is a recursive use of comp with a different parameter B. Beside that technical wrinkle, we see the simplest possible definition of a monad, via a type whose two constructors are precisely the monad operators.
It is easy to define the semantics of terminating comp computations.
Inductive exec A : comp A -> A -> Prop :=
| ExecRet : forall x, exec (Ret x) x
| ExecBnd : forall B (c : comp B) (f : B -> comp A) x1 x2, exec (A := B) c x1
-> exec (f x1) x2
-> exec (Bnd c f) x2.
We can also prove that Ret and Bnd form a monad according to a notion of comp equality based on exec, but we omit details here; they are in the book source at this point.
Not only can we define the Fibonacci function with the new monad, but even our running example of merge sort becomes definable. By shadowing our previous notation for "bind," we can write almost exactly the same code as in our previous mergeSort' definition, but with less syntactic clutter.
Notation "x <- m1 ; m2" := (Bnd m1 (fun x => m2)).
CoFixpoint mergeSort'' A (le : A -> A -> bool) (ls : list A) : comp (list A) :=
if le_lt_dec 2 (length ls)
then let lss := split ls in
ls1 <- mergeSort'' le (fst lss);
ls2 <- mergeSort'' le (snd lss);
Ret (merge le ls1 ls2)
else Ret ls.
To execute this function, we go through the usual exercise of writing a function to catalyze evaluation of co-recursive calls.
Definition frob' A (c : comp A) :=
match c with
| Ret x => Ret x
| Bnd _ c' f => Bnd c' f
end.
Lemma exec_frob : forall A (c : comp A) x,
exec (frob' c) x
-> exec c x.
destruct c; crush.
Qed.
Now the same sort of proof script that we applied for testing thunks will get the job done.
Lemma test_mergeSort'' : exec (mergeSort'' leb (1 :: 2 :: 36 :: 8 :: 19 :: nil))
(1 :: 2 :: 8 :: 19 :: 36 :: nil).
repeat (apply exec_frob; simpl; econstructor).
Qed.
Have we finally reached the ideal solution for encoding general recursive definitions, with minimal hassle in syntax and proof obligations? Unfortunately, we have not, as comp has a serious expressivity weakness. Consider the following definition of a curried addition function:
This definition works fine, but we run into trouble when we try to apply it in a trivial way.
Definition testCurriedAdd := Bnd (curriedAdd 2) (fun f => f 3).
The problem has to do with rules for inductive definitions that we will study in more detail in Chapter 12. Briefly, recall that the type of the constructor Bnd quantifies over a type B. To make testCurriedAdd work, we would need to instantiate B as nat -> comp nat. However, Coq enforces a that (roughly) no quantifier in an inductive or co-inductive type's definition may ever be instantiated with a term that contains the type being defined. Chapter 12 presents the exact mechanism by which this restriction is enforced, but for now our conclusion is that comp is fatally flawed as a way of encoding interesting higher-order functional programs that use general recursion.
We have seen four different approaches to encoding general recursive definitions in Coq. Among them there is no clear champion that dominates the others in every important way. Instead, we close the chapter by comparing the techniques along a number of dimensions. Every technique allows recursive definitions with termination arguments that go beyond Coq's built-in termination checking, so we must turn to subtler points to highlight differences.
One useful property is automatic integration with normal Coq programming. That is, we would like the type of a function to be the same, whether or not that function is defined using an interesting recursion pattern. Only the first of the four techniques, well-founded recursion, meets this criterion. It is also the only one of the four to meet the related criterion that evaluation of function calls can take place entirely inside Coq's built-in computation machinery. The monad inspired by domain theory occupies some middle ground in this dimension, since generally standard computation is enough to evaluate a term once a high enough approximation level is provided.
Another useful property is that a function and its termination argument may be developed separately. We may even want to define functions that fail to terminate on some or all inputs. The well-founded recursion technique does not have this property, but the other three do.
One minor plus is the ability to write recursive definitions in natural syntax, rather than with calls to higher-order combinators. This downside of the first two techniques is actually rather easy to get around using Coq's notation mechanism, though we leave the details as an exercise for the reader. (For this and other details of notations, see Chapter 12 of the Coq 8.4 manual.)
The first two techniques impose proof obligations that are more basic than termination arguments, where well-founded recursion requires a proof of extensionality and domain-theoretic recursion requires a proof of continuity. A function may not be defined, and thus may not be computed with, until these obligations are proved. The co-inductive techniques avoid this problem, as recursive definitions may be made without any proof obligations.
We can also consider support for common idioms in functional programming. For instance, the thunk monad effectively only supports recursion that is tail recursion, while the others allow arbitrary recursion schemes.
On the other hand, the comp monad does not support the effective mixing of higher-order functions and general recursion, while all the other techniques do. For instance, we can finish the failed curriedAdd example in the domain-theoretic monad.
Definition testCurriedAdd := Bnd (curriedAdd 2) (fun f => f 3).
Error: Universe inconsistency.
Comparing the Alternatives
Definition curriedAdd' (n : nat) := Return (fun m : nat => Return (n + m)).
Definition testCurriedAdd := Bind (curriedAdd' 2) (fun f => f 3).
The same techniques also apply to more interesting higher-order functions like list map, and, as in all four techniques, we can mix primitive and general recursion, preferring the former when possible to avoid proof obligations.
Fixpoint map A B (f : A -> computation B) (ls : list A) : computation (list B) :=
match ls with
| nil => Return nil
| x :: ls' => Bind (f x) (fun x' =>
Bind (map f ls') (fun ls'' =>
Return (x' :: ls'')))
end.
Theorem test_map : run (map (fun x => Return (S x)) (1 :: 2 :: 3 :: nil))
(2 :: 3 :: 4 :: nil).
exists 1; reflexivity.
Qed.
One further disadvantage of comp is that we cannot prove an inversion lemma for executions of Bind without appealing to an axiom, a logical complication that we discuss at more length in Chapter 12. The other three techniques allow proof of all the important theorems within the normal logic of Coq.
Perhaps one theme of our comparison is that one must trade off between, on one hand, functional programming expressiveness and compatibility with normal Coq types and computation; and, on the other hand, the level of proof obligations one is willing to handle at function definition time.