-
Notifications
You must be signed in to change notification settings - Fork 2
/
Copy pathchap5.texi
6674 lines (5734 loc) · 284 KB
/
chap5.texi
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
@chapter Computing with Register Machines
@quotation
My aim is to show that the heavenly machine is not a kind of divine, live
being, but a kind of clockwork (and he who believes that a clock has soul
attributes the maker's glory to the work), insofar as nearly all the manifold
motions are caused by a most simple and material force, just as all motions of
the clock are caused by a single weight.
Johannes Kepler (letter to Herwart von Hohenburg, 1605)
@end quotation
We began this book by studying processes and by describing processes in terms
of procedures written in Lisp. To explain the meanings of these procedures, we
used a succession of models of evaluation: the substitution model of
@ref{Chapter 1}, the environment model of @ref{Chapter 3}, and the metacircular
evaluator of @ref{Chapter 4}. Our examination of the metacircular evaluator,
in particular, dispelled much of the mystery of how Lisp-like languages are
interpreted. But even the metacircular evaluator leaves important questions
unanswered, because it fails to elucidate the mechanisms of control in a Lisp
system. For instance, the evaluator does not explain how the evaluation of a
subexpression manages to return a value to the expression that uses this value,
nor does the evaluator explain how some recursive procedures generate iterative
processes (that is, are evaluated using constant space) whereas other recursive
procedures generate recursive processes. These questions remain unanswered
because the metacircular evaluator is itself a Lisp program and hence inherits
the control structure of the underlying Lisp system. In order to provide a
more complete description of the control structure of the Lisp evaluator, we
must work at a more primitive level than Lisp itself.
In this chapter we will describe processes in terms of the step-by-step
operation of a traditional computer. Such a computer, or @newterm{register
machine}, sequentially executes @newterm{instructions} that manipulate the
contents of a fixed set of storage elements called @newterm{registers}. A
typical register-machine instruction applies a primitive operation to the
contents of some registers and assigns the result to another register. Our
descriptions of processes executed by register machines will look very much
like ``machine-language'' programs for traditional computers. However, instead
of focusing on the machine language of any particular computer, we will examine
several Lisp procedures and design a specific register machine to execute each
procedure. Thus, we will approach our task from the perspective of a hardware
architect rather than that of a machine-language computer programmer. In
designing register machines, we will develop mechanisms for implementing
important programming constructs such as recursion. We will also present a
language for describing designs for register machines. In section @ref{5-2} we
will implement a Lisp program that uses these descriptions to simulate the
machines we design.
Most of the primitive operations of our register machines are very simple. For
example, an operation might add the numbers fetched from two registers,
producing a result to be stored into a third register. Such an operation can
be performed by easily described hardware. In order to deal with list
structure, however, we will also use the memory operations @code{car},
@code{cdr}, and @code{cons}, which require an elaborate storage-allocation
mechanism. In section @ref{5-3} we study their implementation in terms of more
elementary operations.
In section @ref{5-4}, after we have accumulated experience formulating simple
procedures as register machines, we will design a machine that carries out the
algorithm described by the metacircular evaluator of section @ref{4-1}. This
will fill in the gap in our understanding of how Scheme expressions are
interpreted, by providing an explicit model for the mechanisms of control in
the evaluator. In section @ref{5-5} we will study a simple compiler that
translates Scheme programs into sequences of instructions that can be executed
directly with the registers and operations of the evaluator register machine.
@menu
* 5-1:: Designing Register Machines
* 5-2:: A Register-Machine Simulator
* 5-3:: Storage Allocation and Garbage Collection
* 5-4:: The Explicit-Control Evaluator
* 5-5:: Compilation
@end menu
@node 5-1, 5-2, Chapter 5, Chapter 5
@section Designing Register Machines
To design a register machine, we must design its @newterm{data paths}
(registers and operations) and the @newterm{controller} that sequences these
operations. To illustrate the design of a simple register machine, let us
examine Euclid's Algorithm, which is used to compute the greatest common
divisor (@acronym{GCD}) of two integers. As we saw in section @ref{1-2-5},
Euclid's Algorithm can be carried out by an iterative process, as specified by
the following procedure:
@lisp
(define (gcd a b)
(if (= b 0)
a
(gcd b (remainder a b))))
@end lisp
A machine to carry out this algorithm must keep track of two numbers, @i{a} and
@i{b}, so let us assume that these numbers are stored in two registers with
those names. The basic operations required are testing whether the contents of
register @code{b} is zero and computing the remainder of the contents of
register @code{a} divided by the contents of register @code{b}. The remainder
operation is a complex process, but assume for the moment that we have a
primitive device that computes remainders. On each cycle of the @acronym{GCD}
algorithm, the contents of register @code{a} must be replaced by the contents
of register @code{b}, and the contents of @code{b} must be replaced by the
remainder of the old contents of @code{a} divided by the old contents of
@code{b}. It would be convenient if these replacements could be done
simultaneously, but in our model of register machines we will assume that only
one register can be assigned a new value at each step. To accomplish the
replacements, our machine will use a third ``temporary'' register, which we
call @code{t}. (First the remainder will be placed in @code{t}, then the
contents of @code{b} will be placed in @code{a}, and finally the remainder
stored in @code{t} will be placed in @code{b}.)
We can illustrate the registers and operations required for this machine by
using the data-path diagram shown in @ref{Figure 5-1}. In this diagram, the
registers (@code{a}, @code{b}, and @code{t}) are represented by rectangles.
Each way to assign a value to a register is indicated by an arrow with an
@code{X} behind the head, pointing from the source of data to the register. We
can think of the @code{X} as a button that, when pushed, allows the value at
the source to ``flow'' into the designated register. The label next to each
button is the name we will use to refer to the button. The names are
arbitrary, and can be chosen to have mnemonic value (for example, @code{a<-b}
denotes pushing the button that assigns the contents of register @code{b} to
register @code{a}). The source of data for a register can be another register
(as in the @code{a<-b} assignment), an operation result (as in the @code{t<-r}
assignment), or a constant (a built-in value that cannot be changed,
represented in a data-path diagram by a triangle containing the constant).
An operation that computes a value from constants and the contents of registers
is represented in a data-path diagram by a trapezoid containing a name for the
operation. For example, the box marked @code{rem} in @ref{Figure 5-1}
represents an operation that computes the remainder of the contents of the
registers @code{a} and @code{b} to which it is attached. Arrows (without
buttons) point from the input registers and constants to the box, and arrows
connect the operation's output value to registers. A test is represented by a
circle containing a name for the test. For example, our @acronym{GCD} machine
has an operation that tests whether the contents of register @code{b} is zero.
A test also has arrows from its input registers and constants, but it has no
output arrows; its value is used by the controller rather than by the data
paths. Overall, the data-path diagram shows the registers and operations that
are required for the machine and how they must be connected. If we view the
arrows as wires and the @code{X} buttons as switches, the data-path diagram is
very like the wiring diagram for a machine that could be constructed from
electrical components.
@quotation
@strong{@anchor{Figure 5-1}Figure 5.1:} Data paths for a @acronym{GCD} machine.
@example
___
+-----+ +-----+ / \
| a |<--(X)----| b +--->| = |
+--+--+ a<-b +-+---+ \___/
| | ^ ^
+------+ +----+ | |
| | (X) b<-t |
.--+---+--. | / \
\ rem / | / O \
\_____/ | +-----+
| |
(X) t<-r |
| |
V |
+-----+ |
| t +------+
+-----+
@end example
@end quotation
In order for the data paths to actually compute @acronym{GCD}s, the buttons
must be pushed in the correct sequence. We will describe this sequence in
terms of a controller diagram, as illustrated in @ref{Figure 5-2}. The
elements of the controller diagram indicate how the data-path components should
be operated. The rectangular boxes in the controller diagram identify
data-path buttons to be pushed, and the arrows describe the sequencing from one
step to the next. The diamond in the diagram represents a decision. One of
the two sequencing arrows will be followed, depending on the value of the
data-path test identified in the diamond. We can interpret the controller in
terms of a physical analogy: Think of the diagram as a maze in which a marble
is rolling. When the marble rolls into a box, it pushes the data-path button
that is named by the box. When the marble rolls into a decision node (such as
the test for @code{b} = 0), it leaves the node on the path determined by the
result of the indicated test. Taken together, the data paths and the
controller completely describe a machine for computing @acronym{GCD}s. We
start the controller (the rolling marble) at the place marked @code{start},
after placing numbers in registers @code{a} and @code{b}. When the controller
reaches @code{done}, we will find the value of the @acronym{GCD} in register
@code{a}.
@quotation
@strong{@anchor{Figure 5-2}Figure 5.2:} Controller for a @acronym{GCD} machine.
@example
start
|
V
/ \ yes
+--->< = >-----> done
| \ /
| | no
| V
| +------+
| | t<-r |
| +---+--+
| |
| V
| +------+
| | a<-b |
| +---+--+
| |
| V
| +------+
+--+ b<-t |
+------+
@end example
@end quotation
@quotation
@strong{@anchor{Exercise 5-1}Exercise 5.1:} Design a register machine to
compute factorials using the iterative algorithm specified by the following
procedure. Draw data-path and controller diagrams for this machine.
@lisp
(define (factorial n)
(define (iter product counter)
(if (> counter n)
product
(iter (* counter product)
(+ counter 1))))
(iter 1 1))
@end lisp
@end quotation
@menu
* 5-1-1:: A Language for Describing Register Machines
* 5-1-2:: Abstraction in Machine Design
* 5-1-3:: Subroutines
* 5-1-4:: Using a Stack to Implement Recursion
* 5-1-5:: Instruction Summary
@end menu
@node 5-1-1, 5-1-2, 5-1, 5-1
@subsection A Language for Describing Register Machines
Data-path and controller diagrams are adequate for representing simple machines
such as @acronym{GCD}, but they are unwieldy for describing large machines such
as a Lisp interpreter. To make it possible to deal with complex machines, we
will create a language that presents, in textual form, all the information
given by the data-path and controller diagrams. We will start with a notation
that directly mirrors the diagrams.
We define the data paths of a machine by describing the registers and the
operations. To describe a register, we give it a name and specify the buttons
that control assignment to it. We give each of these buttons a name and
specify the source of the data that enters the register under the button's
control. (The source is a register, a constant, or an operation.) To describe
an operation, we give it a name and specify its inputs (registers or
constants).
We define the controller of a machine as a sequence of @newterm{instructions}
together with @newterm{labels} that identify @newterm{entry points} in the
sequence. An instruction is one of the following:
@itemize @bullet
@item
The name of a data-path button to push to assign a value to a register. (This
corresponds to a box in the controller diagram.)
@item
A @code{test} instruction, that performs a specified test.
@item
A conditional branch (@code{branch} instruction) to a location indicated by a
controller label, based on the result of the previous test. (The test and
branch together correspond to a diamond in the controller diagram.) If the
test is false, the controller should continue with the next instruction in the
sequence. Otherwise, the controller should continue with the instruction after
the label.
@item
An unconditional branch (@code{goto} instruction) naming a controller label at
which to continue execution.
@end itemize
The machine starts at the beginning of the controller instruction sequence and
stops when execution reaches the end of the sequence. Except when a branch
changes the flow of control, instructions are executed in the order in which
they are listed.
@quotation
@strong{@anchor{Figure 5-3}Figure 5.3:} A specification of the @acronym{GCD}
machine.
@lisp
(data-paths
(registers
((name a)
(buttons ((name a<-b) (source (register b)))))
((name b)
(buttons ((name b<-t) (source (register t)))))
((name t)
(buttons ((name t<-r) (source (operation rem))))))
(operations
((name rem)
(inputs (register a) (register b)))
((name =)
(inputs (register b) (constant 0)))))
(controller
test-b @r{; label}
(test =) @r{; test}
(branch (label gcd-done)) @r{; conditional branch}
(t<-r) @r{; button push}
(a<-b) @r{; button push}
(b<-t) @r{; button push}
(goto (label test-b)) @r{; unconditional branch}
gcd-done) @r{; label}
@end lisp
@end quotation
@ref{Figure 5-3} shows the @acronym{GCD} machine described in this way. This
example only hints at the generality of these descriptions, since the
@acronym{GCD} machine is a very simple case: Each register has only one button,
and each button and test is used only once in the controller.
Unfortunately, it is difficult to read such a description. In order to
understand the controller instructions we must constantly refer back to the
definitions of the button names and the operation names, and to understand what
the buttons do we may have to refer to the definitions of the operation names.
We will thus transform our notation to combine the information from the
data-path and controller descriptions so that we see it all together.
To obtain this form of description, we will replace the arbitrary button and
operation names by the definitions of their behavior. That is, instead of
saying (in the controller) ``Push button @code{t<-r}'' and separately saying
(in the data paths) ``Button @code{t<-r} assigns the value of the @code{rem}
operation to register @code{t}'' and ``The @code{rem} operation's inputs are
the contents of registers @code{a} and @code{b},'' we will say (in the
controller) ``Push the button that assigns to register @code{t} the value of
the @code{rem} operation on the contents of registers @code{a} and @code{b}.''
Similarly, instead of saying (in the controller) ``Perform the @code{=} test''
and separately saying (in the data paths) ``The @code{=} test operates on the
contents of register @code{b} and the constant 0,'' we will say ``Perform the
@code{=} test on the contents of register @code{b} and the constant 0.'' We
will omit the data-path description, leaving only the controller sequence.
Thus, the @acronym{GCD} machine is described as follows:
@lisp
(controller
test-b
(test (op =) (reg b) (const 0))
(branch (label gcd-done))
(assign t (op rem) (reg a) (reg b))
(assign a (reg b))
(assign b (reg t))
(goto (label test-b))
gcd-done)
@end lisp
This form of description is easier to read than the kind illustrated in
@ref{Figure 5-3}, but it also has disadvantages:
@itemize @bullet
@item
It is more verbose for large machines, because complete descriptions of the
data-path elements are repeated whenever the elements are mentioned in the
controller instruction sequence. (This is not a problem in the @acronym{GCD}
example, because each operation and button is used only once.) Moreover,
repeating the data-path descriptions obscures the actual data-path structure of
the machine; it is not obvious for a large machine how many registers,
operations, and buttons there are and how they are interconnected.
@item
Because the controller instructions in a machine definition look like Lisp
expressions, it is easy to forget that they are not arbitrary Lisp expressions.
They can notate only legal machine operations. For example, operations can
operate directly only on constants and the contents of registers, not on the
results of other operations.
@end itemize
In spite of these disadvantages, we will use this register-machine language
throughout this chapter, because we will be more concerned with understanding
controllers than with understanding the elements and connections in data paths.
We should keep in mind, however, that data-path design is crucial in designing
real machines.
@quotation
@strong{@anchor{Exercise 5-2}Exercise 5.2:} Use the register-machine language
to describe the iterative factorial machine of @ref{Exercise 5-1}.
@end quotation
@subsubheading Actions
Let us modify the @acronym{GCD} machine so that we can type in the numbers
whose @acronym{GCD} we want and get the answer printed at our terminal. We
will not discuss how to make a machine that can read and print, but will assume
(as we do when we use @code{read} and @code{display} in Scheme) that they are
available as primitive operations.@footnote{This assumption glosses over a
great deal of complexity. Usually a large portion of the implementation of a
Lisp system is dedicated to making reading and printing work.}
@code{Read} is like the operations we have been using in that it produces a
value that can be stored in a register. But @code{read} does not take inputs
from any registers; its value depends on something that happens outside the
parts of the machine we are designing. We will allow our machine's operations
to have such behavior, and thus will draw and notate the use of @code{read}
just as we do any other operation that computes a value.
@code{Print}, on the other hand, differs from the operations we have been using
in a fundamental way: It does not produce an output value to be stored in a
register. Though it has an effect, this effect is not on a part of the machine
we are designing. We will refer to this kind of operation as an
@newterm{action}. We will represent an action in a data-path diagram just as
we represent an operation that computes a value---as a trapezoid that contains
the name of the action. Arrows point to the action box from any inputs
(registers or constants). We also associate a button with the action. Pushing
the button makes the action happen. To make a controller push an action button
we use a new kind of instruction called @code{perform}. Thus, the action of
printing the contents of register @code{a} is represented in a controller
sequence by the instruction
@lisp
(perform (op print) (reg a))
@end lisp
@ref{Figure 5-4} shows the data paths and controller for the new @acronym{GCD}
machine. Instead of having the machine stop after printing the answer, we have
made it start over, so that it repeatedly reads a pair of numbers, computes
their @acronym{GCD}, and prints the result. This structure is like the driver
loops we used in the interpreters of @ref{Chapter 4}.
@quotation
@strong{@anchor{Figure 5-4}Figure 5.4:} A @acronym{GCD} machine that reads
inputs and prints results.
@example
.--------.
\ read /
\____/
|
+-------*------+
| |
a<-rd (X) (X) b<-rd
| |
V V ___
+-----+ +-----+ / \
| a |<--(X)--+ b +----->| = |
+-+-+-+ a<-b +-+---+ \___/
| | | ^ ^
+--+ +----+ +--+ | |
| | | (X) b<-t / \
V V V | / O \
.---------. .---------. | /_____\
--(X)->\ print / \ rem / |
P \_____/ \_____/ |
| |
(X) t<-r |
| |
V |
+-----+ |
| t +----+
+-----+
@end example
@lisp
(controller
gcd-loop
(assign a (op read))
(assign b (op read))
test-b
(test (op =) (reg b) (const 0))
(branch (label gcd-done))
(assign t (op rem) (reg a) (reg b))
(assign a (reg b))
(assign b (reg t))
(goto (label test-b))
gcd-done
(perform (op print) (reg a))
(goto (label gcd-loop)))
@end lisp
@end quotation
@node 5-1-2, 5-1-3, 5-1-1, 5-1
@subsection Abstraction in Machine Design
We will often define a machine to include ``primitive'' operations that are
actually very complex. For example, in sections @ref{5-4} and @ref{5-5} we
will treat Scheme's environment manipulations as primitive. Such abstraction
is valuable because it allows us to ignore the details of parts of a machine so
that we can concentrate on other aspects of the design. The fact that we have
swept a lot of complexity under the rug, however, does not mean that a machine
design is unrealistic. We can always replace the complex ``primitives'' by
simpler primitive operations.
Consider the @acronym{GCD} machine. The machine has an instruction that
computes the remainder of the contents of registers @code{a} and @code{b} and
assigns the result to register @code{t}. If we want to construct the
@acronym{GCD} machine without using a primitive remainder operation, we must
specify how to compute remainders in terms of simpler operations, such as
subtraction. Indeed, we can write a Scheme procedure that finds remainders in
this way:
@lisp
(define (remainder n d)
(if (< n d)
n
(remainder (- n d) d)))
@end lisp
We can thus replace the remainder operation in the @acronym{GCD} machine's data
paths with a subtraction operation and a comparison test. @ref{Figure 5-5}
shows the data paths and controller for the elaborated machine. The
instruction
@quotation
@strong{@anchor{Figure 5-5}Figure 5.5:} Data paths and controller for the
elaborated @acronym{GCD} machine.
@example
___
+-----+ +-----+ / \
| a |<--(X)---+ b +-------*-->| = |
+--+--+ a<-b +-+---+ | \___/
| | ^ |
(X) t<-a | | |
| | (X) b<-t |
V | | _V_
+-----+ | | / \
| t +-------*---|--*-----| < |
+-----+ | | \___/
^ V V
| ---------
(X) t<-d \ - /
| --+--
| |
+------------+
start
|
V
/ \ yes +-------+
+->< = >----> done | t<-d |<--+
| \ / +---+---+ |
| | no | |
| | V |
| | +------+ / \ no |
| +-->| t<-a +------->< < >-----+
| +------+ \ /
| | yes
| +-------------------+
| V
| +-------+
| | a<-b |
| +---+---+
| |
| V
| +-------+
+--+ b<-t |
+-------+
@end example
@end quotation
@lisp
(assign t (op rem) (reg a) (reg b))
@end lisp
@noindent
in the @acronym{GCD} controller definition is replaced by a sequence of
instructions that contains a loop, as shown in @ref{Figure 5-6}.
@quotation
@strong{@anchor{Figure 5-6}Figure 5.6:} Controller instruction sequence for the
@acronym{GCD} machine in @ref{Figure 5-5}.
@lisp
(controller
test-b
(test (op =) (reg b) (const 0))
(branch (label gcd-done))
(assign t (reg a))
rem-loop
(test (op <) (reg t) (reg b))
(branch (label rem-done))
(assign t (op -) (reg t) (reg b))
(goto (label rem-loop))
rem-done
(assign a (reg b))
(assign b (reg t))
(goto (label test-b))
gcd-done)
@end lisp
@end quotation
@quotation
@strong{@anchor{Exercise 5-3}Exercise 5.3:} Design a machine to compute square
roots using Newton's method, as described in section @ref{1-1-7}:
@lisp
(define (sqrt x)
(define (good-enough? guess)
(< (abs (- (square guess) x)) 0.001))
(define (improve guess)
(average guess (/ x guess)))
(define (sqrt-iter guess)
(if (good-enough? guess)
guess
(sqrt-iter (improve guess))))
(sqrt-iter 1.0))
@end lisp
Begin by assuming that @code{good-enough?} and @code{improve} operations are
available as primitives. Then show how to expand these in terms of arithmetic
operations. Describe each version of the @code{sqrt} machine design by drawing
a data-path diagram and writing a controller definition in the register-machine
language.
@end quotation
@node 5-1-3, 5-1-4, 5-1-2, 5-1
@subsection Subroutines
When designing a machine to perform a computation, we would often prefer to
arrange for components to be shared by different parts of the computation
rather than duplicate the components. Consider a machine that includes two
@acronym{GCD} computations---one that finds the @acronym{GCD} of the contents
of registers @code{a} and @code{b} and one that finds the @acronym{GCD} of the
contents of registers @code{c} and @code{d}. We might start by assuming we
have a primitive @code{gcd} operation, then expand the two instances of
@code{gcd} in terms of more primitive operations. @ref{Figure 5-7} shows just
the @acronym{GCD} portions of the resulting machine's data paths, without
showing how they connect to the rest of the machine. The figure also shows the
corresponding portions of the machine's controller sequence.
@quotation
@strong{@anchor{Figure 5-7}Figure 5.7:} Portions of the data paths and
controller sequence for a machine with two @acronym{GCD} computations.
@example
___ ___
+-----+ +-----+ / \ +-----+ +-----+ / \
| a |<-(X)---+ b |--->| = | | c |<-(X)---+ d |--->| = |
+--+--+ a<-b ++----+ \___/ +--+--+ c<-d ++----+ \___/
| | ^ ^ | | ^ ^
`----. .---' | | `----. .---' | |
V V (X) b<-t | V V (X) d<-t |
------- | / \ ------- | / \
\ rem / | /_0_\ \ rem / | /_0_\
--+-- | --+-- |
| | | |
(X) t<-r | (X) s<-r |
| | | |
V | V |
+-----+ | +-----+ |
| t +-----' | s +-----'
+-----+ +-----+
@end example
@lisp
gcd-1
(test (op =) (reg b) (const 0))
(branch (label after-gcd-1))
(assign t (op rem) (reg a) (reg b))
(assign a (reg b))
(assign b (reg t))
(goto (label gcd-1))
after-gcd-1
@dots{}
gcd-2
(test (op =) (reg d) (const 0))
(branch (label after-gcd-2))
(assign s (op rem) (reg c) (reg d))
(assign c (reg d))
(assign d (reg s))
(goto (label gcd-2))
after-gcd-2
@end lisp
@end quotation
This machine has two remainder operation boxes and two boxes for testing
equality. If the duplicated components are complicated, as is the remainder
box, this will not be an economical way to build the machine. We can avoid
duplicating the data-path components by using the same components for both
@acronym{GCD} computations, provided that doing so will not affect the rest of
the larger machine's computation. If the values in registers @code{a} and
@code{b} are not needed by the time the controller gets to @code{gcd-2} (or if
these values can be moved to other registers for safekeeping), we can change
the machine so that it uses registers @code{a} and @code{b}, rather than
registers @code{c} and @code{d}, in computing the second @acronym{GCD} as well
as the first. If we do this, we obtain the controller sequence shown in
@ref{Figure 5-8}.
We have removed the duplicate data-path components (so that the data paths are
again as in @ref{Figure 5-1}), but the controller now has two @acronym{GCD}
sequences that differ only in their entry-point labels. It would be better to
replace these two sequences by branches to a single sequence---a @code{gcd}
@newterm{subroutine}---at the end of which we branch back to the correct place
in the main instruction sequence. We can accomplish this as follows: Before
branching to @code{gcd}, we place a distinguishing value (such as 0 or 1) into
a special register, @code{continue}. At the end of the @code{gcd} subroutine
we return either to @code{after-gcd-1} or to @code{after-gcd-2}, depending on
the value of the @code{continue} register. @ref{Figure 5-9} shows the relevant
portion of the resulting controller sequence, which includes only a single copy
of the @code{gcd} instructions.
@quotation
@strong{@anchor{Figure 5-8}Figure 5.8:} Portions of the controller sequence for
a machine that uses the same data-path components for two different
@acronym{GCD} computations.
@lisp
gcd-1
(test (op =) (reg b) (const 0))
(branch (label after-gcd-1))
(assign t (op rem) (reg a) (reg b))
(assign a (reg b))
(assign b (reg t))
(goto (label gcd-1))
after-gcd-1
@dots{}
gcd-2
(test (op =) (reg b) (const 0))
(branch (label after-gcd-2))
(assign t (op rem) (reg a) (reg b))
(assign a (reg b))
(assign b (reg t))
(goto (label gcd-2))
after-gcd-2
@end lisp
@end quotation
@quotation
@strong{@anchor{Figure 5-9}Figure 5.9:} Using a @code{continue} register to
avoid the duplicate controller sequence in @ref{Figure 5-8}.
@lisp
gcd
(test (op =) (reg b) (const 0))
(branch (label gcd-done))
(assign t (op rem) (reg a) (reg b))
(assign a (reg b))
(assign b (reg t))
(goto (label gcd))
gcd-done
(test (op =) (reg continue) (const 0))
(branch (label after-gcd-1))
(goto (label after-gcd-2))
@dots{}
@r{;; Before branching to @code{gcd} from the first place where}
@r{;; it is needed, we place 0 in the @code{continue} register}
(assign continue (const 0))
(goto (label gcd))
after-gcd-1
@dots{}
@r{;; Before the second use of @code{gcd}, we place 1 in the @code{continue} register}
(assign continue (const 1))
(goto (label gcd))
after-gcd-2
@end lisp
@end quotation
@quotation
@strong{@anchor{Figure 5-10}Figure 5.10:} Assigning labels to the
@code{continue} register simplifies and generalizes the strategy shown in
@ref{Figure 5-9}.
@lisp
gcd
(test (op =) (reg b) (const 0))
(branch (label gcd-done))
(assign t (op rem) (reg a) (reg b))
(assign a (reg b))
(assign b (reg t))
(goto (label gcd))
gcd-done
(goto (reg continue))
@dots{}
@r{;; Before calling @code{gcd}, we assign to @code{continue}}
@r{;; the label to which @code{gcd} should return.}
(assign continue (label after-gcd-1))
(goto (label gcd))
after-gcd-1
@dots{}
@r{;; Here is the second call to @code{gcd}, with a different continuation.}
(assign continue (label after-gcd-2))
(goto (label gcd))
after-gcd-2
@end lisp
@end quotation
This is a reasonable approach for handling small problems, but it would be
awkward if there were many instances of @acronym{GCD} computations in the
controller sequence. To decide where to continue executing after the
@code{gcd} subroutine, we would need tests in the data paths and branch
instructions in the controller for all the places that use @code{gcd}. A more
powerful method for implementing subroutines is to have the @code{continue}
register hold the label of the entry point in the controller sequence at which
execution should continue when the subroutine is finished. Implementing this
strategy requires a new kind of connection between the data paths and the
controller of a register machine: There must be a way to assign to a register a
label in the controller sequence in such a way that this value can be fetched
from the register and used to continue execution at the designated entry point.
To reflect this ability, we will extend the @code{assign} instruction of the
register-machine language to allow a register to be assigned as value a label
from the controller sequence (as a special kind of constant). We will also
extend the @code{goto} instruction to allow execution to continue at the entry
point described by the contents of a register rather than only at an entry
point described by a constant label. Using these new constructs we can
terminate the @code{gcd} subroutine with a branch to the location stored in the
@code{continue} register. This leads to the controller sequence shown in
@ref{Figure 5-10}.
A machine with more than one subroutine could use multiple continuation
registers (e.g., @code{gcd-continue}, @code{factorial-continue}) or we could
have all subroutines share a single @code{continue} register. Sharing is more
economical, but we must be careful if we have a subroutine (@code{sub1}) that
calls another subroutine (@code{sub2}). Unless @code{sub1} saves the contents
of @code{continue} in some other register before setting up @code{continue} for
the call to @code{sub2}, @code{sub1} will not know where to go when it is
finished. The mechanism developed in the next section to handle recursion also
provides a better solution to this problem of nested subroutine calls.
@node 5-1-4, 5-1-5, 5-1-3, 5-1
@subsection Using a Stack to Implement Recursion
With the ideas illustrated so far, we can implement any iterative process by
specifying a register machine that has a register corresponding to each state
variable of the process. The machine repeatedly executes a controller loop,
changing the contents of the registers, until some termination condition is
satisfied. At each point in the controller sequence, the state of the machine
(representing the state of the iterative process) is completely determined by
the contents of the registers (the values of the state variables).
Implementing recursive processes, however, requires an additional mechanism.
Consider the following recursive method for computing factorials, which we
first examined in section @ref{1-2-1}:
@lisp
(define (factorial n)
(if (= n 1)
1
(* (factorial (- n 1)) n)))
@end lisp
As we see from the procedure, computing @i{n}! requires computing (@i{n} - 1)!.
Our @acronym{GCD} machine, modeled on the procedure
@lisp
(define (gcd a b)
(if (= b 0)
a
(gcd b (remainder a b))))
@end lisp
@noindent
similarly had to compute another @acronym{GCD}. But there is an important
difference between the @code{gcd} procedure, which reduces the original
computation to a new @acronym{GCD} computation, and @code{factorial}, which
requires computing another factorial as a subproblem. In @acronym{GCD}, the
answer to the new @acronym{GCD} computation is the answer to the original
problem. To compute the next @acronym{GCD}, we simply place the new arguments
in the input registers of the @acronym{GCD} machine and reuse the machine's
data paths by executing the same controller sequence. When the machine is
finished solving the final @acronym{GCD} problem, it has completed the entire
computation.
In the case of factorial (or any recursive process) the answer to the new
factorial subproblem is not the answer to the original problem. The value
obtained for (@i{n} - 1)! must be multiplied by @i{n} to get the final answer.
If we try to imitate the @acronym{GCD} design, and solve the factorial
subproblem by decrementing the @code{n} register and rerunning the factorial
machine, we will no longer have available the old value of @code{n} by which to
multiply the result. We thus need a second factorial machine to work on the
subproblem. This second factorial computation itself has a factorial
subproblem, which requires a third factorial machine, and so on. Since each
factorial machine contains another factorial machine within it, the total
machine contains an infinite nest of similar machines and hence cannot be
constructed from a fixed, finite number of parts.
Nevertheless, we can implement the factorial process as a register machine if
we can arrange to use the same components for each nested instance of the
machine. Specifically, the machine that computes @i{n}! should use the same
components to work on the subproblem of computing (@i{n} - 1)!, on the
subproblem for (@i{n} - 2)!, and so on. This is plausible because, although
the factorial process dictates that an unbounded number of copies of the same
machine are needed to perform a computation, only one of these copies needs to
be active at any given time. When the machine encounters a recursive
subproblem, it can suspend work on the main problem, reuse the same physical
parts to work on the subproblem, then continue the suspended computation.
In the subproblem, the contents of the registers will be different than they
were in the main problem. (In this case the @code{n} register is decremented.)
In order to be able to continue the suspended computation, the machine must
save the contents of any registers that will be needed after the subproblem is
solved so that these can be restored to continue the suspended computation. In
the case of factorial, we will save the old value of @code{n}, to be restored
when we are finished computing the factorial of the decremented @code{n}
register.@footnote{One might argue that we don't need to save the old @code{n};
after we decrement it and solve the subproblem, we could simply increment it to
recover the old value. Although this strategy works for factorial, it cannot
work in general, since the old value of a register cannot always be computed
from the new one.}
Since there is no @emph{a priori} limit on the depth of nested recursive calls,
we may need to save an arbitrary number of register values. These values must
be restored in the reverse of the order in which they were saved, since in a
nest of recursions the last subproblem to be entered is the first to be
finished. This dictates the use of a @newterm{stack}, or ``last in, first
out'' data structure, to save register values. We can extend the
register-machine language to include a stack by adding two kinds of
instructions: Values are placed on the stack using a @code{save} instruction
and restored from the stack using a @code{restore} instruction. After a
sequence of values has been @code{save}d on the stack, a sequence of
@code{restore}s will retrieve these values in reverse order.@footnote{In
section @ref{5-3} we will see how to implement a stack in terms of more
primitive operations.}
With the aid of the stack, we can reuse a single copy of the factorial
machine's data paths for each factorial subproblem. There is a similar design
issue in reusing the controller sequence that operates the data paths. To
reexecute the factorial computation, the controller cannot simply loop back to
the beginning, as with an iterative process, because after solving the (@i{n} -
1)! subproblem the machine must still multiply the result by @i{n}. The
controller must suspend its computation of @i{n}!, solve the (@i{n} - 1)!
subproblem, then continue its computation of @i{n}!. This view of the
factorial computation suggests the use of the subroutine mechanism described in
section @ref{5-1-3}, which has the controller use a @code{continue} register to
transfer to the part of the sequence that solves a subproblem and then continue
where it left off on the main problem. We can thus make a factorial subroutine
that returns to the entry point stored in the @code{continue} register. Around
each subroutine call, we save and restore @code{continue} just as we do the
@code{n} register, since each ``level'' of the factorial computation will use
the same @code{continue} register. That is, the factorial subroutine must put
a new value in @code{continue} when it calls itself for a subproblem, but it
will need the old value in order to return to the place that called it to solve
a subproblem.
@ref{Figure 5-11} shows the data paths and controller for a machine that
implements the recursive @code{factorial} procedure. The machine has a stack
and three registers, called @code{n}, @code{val}, and @code{continue}. To
simplify the data-path diagram, we have not named the register-assignment
buttons, only the stack-operation buttons (@code{sc} and @code{sn} to save
registers, @code{rc} and @code{rn} to restore registers). To operate the
machine, we put in register @code{n} the number whose factorial we wish to
compute and start the machine. When the machine reaches @code{fact-done}, the
computation is finished and the answer will be found in the @code{val}
register. In the controller sequence, @code{n} and @code{continue} are saved
before each recursive call and restored upon return from the call. Returning
from a call is accomplished by branching to the location stored in
@code{continue}. @code{Continue} is initialized when the machine starts so
that the last return will go to @code{fact-done}. The @code{val} register,
which holds the result of the factorial computation, is not saved before the
recursive call, because the old contents of @code{val} is not useful after the
subroutine returns. Only the new value, which is the value produced by the
subcomputation, is needed.
Although in principle the factorial computation requires an infinite machine,
the machine in @ref{Figure 5-11} is actually finite except for the stack, which
is potentially unbounded. Any particular physical implementation of a stack,
however, will be of finite size, and this will limit the depth of recursive
calls that can be handled by the machine. This implementation of factorial
illustrates the general strategy for realizing recursive algorithms as ordinary
register machines augmented by stacks. When a recursive subproblem is
encountered, we save on the stack the registers whose current values will be
required after the subproblem is solved, solve the recursive subproblem, then
restore the saved registers and continue execution on the main problem. The
@code{continue} register must always be saved. Whether there are other
registers that need to be saved depends on the particular machine, since not
all recursive computations need the original values of registers that are
modified during solution of the subproblem (see @ref{Exercise 5-4}).
@subsubheading A double recursion
Let us examine a more complex recursive process, the tree-recursive computation
of the Fibonacci numbers, which we introduced in section @ref{1-2-2}:
@lisp
(define (fib n)
(if (< n 2)
n
(+ (fib (- n 1)) (fib (- n 2)))))
@end lisp
Just as with factorial, we can implement the recursive Fibonacci computation as
a register machine with registers @code{n}, @code{val}, and @code{continue}.
The machine is more complex than the one for factorial, because there are two
places in the controller sequence where we need to perform recursive
calls---once to compute Fib(@i{n} - 1) and once to compute Fib(@i{n} - 2). To
set up for each of these calls, we save the registers whose values will be
needed later, set the @code{n} register to the number whose Fib we need to
compute recursively (@i{n} - 1 or @i{n} - 2), and assign to @code{continue} the