Difference between revisions of "Directory:Jon Awbrey/Papers/Differential Analytic Turing Automata"

MyWikiBiz, Author Your Legacy — Sunday November 24, 2024
Jump to navigationJump to search
(→‎State Partition: <math>\operatorname{M}</math>)
(update)
 
(31 intermediate revisions by the same user not shown)
Line 1: Line 1:
 
{{DISPLAYTITLE:Differential Analytic Turing Automata}}
 
{{DISPLAYTITLE:Differential Analytic Turing Automata}}
 +
'''Author: [[User:Jon Awbrey|Jon Awbrey]]'''
  
==Note 1==
+
The task ahead is to chart a course from general ideas about ''transformational equivalence classes of graphs'' to a notion of ''differential analytic turing automata'' (DATA).  It may be a while before we get within sight of that goal, but it will provide a better measure of motivation to name the thread after the envisioned end rather than the more homely starting place.
  
My aim is to chart a course from general ideas about ''transformational equivalence classes of graphs'' to a notion of ''differential analytic turing automata'' (DATA).  It may be a while before we get within sight of that goal, but it will provide some measure of motivation to name the thread after the envisioned end rather than the more homely starting place.
+
The basic idea is as follows.  One has a set <math>\mathcal{G}</math> of graphs and a set <math>\mathcal{T}</math> of transformation rules, and each rule <math>\mathrm{t} \in \mathcal{T}</math> has the effect of transforming graphs into graphs, <math>\mathrm{t} : \mathcal{G} \to \mathcal{G}.</math>  In the cases that we shall be studying, this set of transformation rules partitions the set of graphs into ''transformational equivalence classes'' (TECs).
 
 
The basic idea here is that you have a species of graphs and a set of transformation rules that take you from one graph to another &mdash; and back again, as I'm only thinking of equational rules &mdash; and this partitions the species of graphs into ''transformational equivalence classes'' (TECs).
 
  
 
There are many interesting excursions to be had here, but I will focus mainly on logical applications, and and so the TECs I talk about will almost always have the character of ''logical equivalence classes'' (LECs).
 
There are many interesting excursions to be had here, but I will focus mainly on logical applications, and and so the TECs I talk about will almost always have the character of ''logical equivalence classes'' (LECs).
  
An example that will figure heavily in the sequel is given by rooted trees as the species of graphs and a pair of equational transformation rules that derive from the graphical calculi of [[C.S. Peirce]], as revived and extended by George Spencer Brown.
+
An example that will figure heavily in the sequel is given by rooted trees as the species of graphs and a pair of equational transformation rules that derive from the graphical calculi of C.S. Peirce, as revived and extended by George Spencer Brown.
  
 
Here are the fundamental transformation rules, also referred to as the ''arithmetic axioms'', more precisely, the ''arithmetic initials''.
 
Here are the fundamental transformation rules, also referred to as the ''arithmetic axioms'', more precisely, the ''arithmetic initials''.
Line 21: Line 20:
 
That should be enough to get started.
 
That should be enough to get started.
  
==Note 2==
+
==Cactus Language==
  
 
I will be making use of the ''cactus language'' extension of Peirce's Alpha Graphs, so called because it uses a species of graphs that are usually called "cacti" in graph theory.  The last exposition of the cactus syntax that I've written can be found here:
 
I will be making use of the ''cactus language'' extension of Peirce's Alpha Graphs, so called because it uses a species of graphs that are usually called "cacti" in graph theory.  The last exposition of the cactus syntax that I've written can be found here:
  
:* [http://mywikibiz.com/Directory:Jon_Awbrey/Papers/Propositional_Equation_Reasoning_Systems Propositional Equation Reasoning Systems (PERS)]
+
:* [[Propositional Equation Reasoning Systems|Propositional Equation Reasoning Systems (PERS)]]
  
 
The representational and computational efficiency of the cactus language for the tasks that are usually associated with boolean algebra and propositional calculus makes it possible to entertain a further extension, to what we may call ''differential logic'', because it develops this basic level of logic in the same way that differential calculus augments analytic geometry to handle change and diversity.  There are several different introductions to differential logic that I have written and distributed across the Internet.  You might start with the following couple of treatments:
 
The representational and computational efficiency of the cactus language for the tasks that are usually associated with boolean algebra and propositional calculus makes it possible to entertain a further extension, to what we may call ''differential logic'', because it develops this basic level of logic in the same way that differential calculus augments analytic geometry to handle change and diversity.  There are several different introductions to differential logic that I have written and distributed across the Internet.  You might start with the following couple of treatments:
Line 32: Line 31:
 
:* [http://stderr.org/pipermail/inquiry/2004-February/thread.html#1160 Differential Logic B]
 
:* [http://stderr.org/pipermail/inquiry/2004-February/thread.html#1160 Differential Logic B]
  
I am currently rewriting these presentations in hopes of making them as clear as they can be, so please let me know if you have any questions.
+
I will draw on those previously advertised resources of notation and theory as needed, but right now I sense the need for some concrete examples.
 
 
==Note 3==
 
  
I will draw on those previously advertised resources of notation and theory as needed, but right now I sense the need for some concrete examples.
+
==Example 1==
  
Let's say we have a system that is known by the name of its state space <math>X\!</math> and we have a boolean state variable <math>x : X \to \mathbb{B},</math> where <math>\mathbb{B} = \{ 0, 1 \}.</math>
+
Let's say we have a system that is known by the name of its state space <math>X\!</math> and we have a boolean state variable <math>x : X \to \mathbb{B},\!</math> where <math>\mathbb{B} = \{ 0, 1 \}.\!</math>
  
 
We observe <math>X\!</math> for a while, relative to a discrete time frame, and we write down the following sequence of values for <math>x.\!</math>
 
We observe <math>X\!</math> for a while, relative to a discrete time frame, and we write down the following sequence of values for <math>x.\!</math>
Line 44: Line 41:
 
{| align="center" cellpadding="8" style="text-align:center"
 
{| align="center" cellpadding="8" style="text-align:center"
 
|
 
|
<math>\begin{array}{cc}
+
<math>\begin{array}{c|c}
t & x \\
+
t & x \\[8pt]
\\
 
 
0 & 0 \\
 
0 & 0 \\
 
1 & 1 \\
 
1 & 1 \\
Line 60: Line 56:
 
|}
 
|}
  
"Aha!" we say, and think we see the way of things, writing down the rule <math>\texttt{x' = (x)}</math> where <math>\texttt{x'}</math> is the state that comes next after <math>\texttt{x},</math> and <math>\texttt{(x)}</math> is the negation of <math>\texttt{x}</math> in boolean logic.
+
&ldquo;Aha!&rdquo; we say, and think we see the way of things, writing down the rule <math>x' = \texttt{(} x \texttt{)},\!</math> where <math>x'\!</math> is the next state after <math>x\!</math> and <math>\texttt{(} x \texttt{)}\!</math> is the negation of <math>x\!</math> in boolean logic.
  
 
Another way to detect patterns is to write out a table of finite differences.  For this example, we would get:
 
Another way to detect patterns is to write out a table of finite differences.  For this example, we would get:
Line 66: Line 62:
 
{| align="center" cellpadding="8" style="text-align:center"
 
{| align="center" cellpadding="8" style="text-align:center"
 
|
 
|
<math>\begin{array}{ccccc}
+
<math>\begin{array}{c|cccc}
t &     x &     dx & d^2 x & \ldots \\
+
t & x & \mathrm{d}x & \mathrm{d}^2 x & \ldots \\[8pt]
\\
+
0 & 0 & 1 & 0 & \ldots \\
0 &     0 &     1 &     0 & \ldots \\
+
1 & 1 & 1 & 0 & \ldots \\
1 &     1 &     1 &     0 &       \\
+
2 & 0 & 1 & 0 & \ldots \\
2 &     0 &     1 &     0 &       \\
+
3 & 1 & 1 & 0 & \ldots \\
3 &     1 &     1 &     0 &       \\
+
4 & 0 & 1 & 0 & \ldots \\
4 &     0 &     1 &     0 &       \\
+
5 & 1 & 1 & 0 & \ldots \\
5 &     1 &     1 &     0 &       \\
+
6 & 0 & 1 & 0 & \ldots \\
6 &     0 &     1 &     0 &       \\
+
7 & 1 & 1 & 0 & \ldots \\
7 &     1 &     1 &     0 &       \\
+
8 & 0 & 1 & \ldots & \ldots \\
8 &     0 &     1 & \ldots &       \\
+
9 & \ldots & \ldots & \ldots & \ldots
9 & \ldots & \ldots & \ldots &       \\
 
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
Line 84: Line 79:
 
And of course, all the higher order differences are zero.
 
And of course, all the higher order differences are zero.
  
This leads to thinking of <math>X\!</math> as having an extended state <math>(x, dx, d^2 x, \ldots, d^k x),</math> and this additional language gives us the facility of describing state transitions in terms of the various orders of differences.  For example, the rule <math>\texttt{x' = (x)}</math> can now be expressed by the rule <math>\texttt{dx = 1}.</math>
+
This leads to thinking of <math>X\!</math> as having an extended state <math>(x, \mathrm{d}x, \mathrm{d}^2 x, \ldots, \mathrm{d}^k x),\!</math> and this additional language gives us the facility of describing state transitions in terms of the various orders of differences.  For example, the rule <math>x' = \texttt{(} x \texttt{)}\!</math> can now be expressed by the rule <math>{\mathrm{d}x = 1}.\!</math>
 
 
I'll leave you to muse on the possibilities of that.
 
 
 
==Note 4==
 
  
I am preparing a more fleshed-out 1-variable example, but in the mean time, for anybody who's finished all that other reading, there is a more detailed account of differential logic in the following paper:
+
There is a more detailed account of differential logic in the following paper:
  
:* [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0|Differential Logic and Dynamic Systems]]
+
:* [[Differential Logic and Dynamic Systems 2.0|Differential Logic and Dynamic Systems]]
  
For future reference, here are a couple of handy rosetta stones for translating back and forth between different notations for the boolean functions <math>f : \mathbb{B}^k \to \mathbb{B},</math> where <math>k = 1, 2.\!</math>
+
For future reference, here are a couple of handy rosetta stones for translating back and forth between different notations for the boolean functions <math>f : \mathbb{B}^k \to \mathbb{B},\!</math> where <math>k = 1, 2.\!</math>
  
:* [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0#Tables_of_Propositional_Forms|Tables of Propositional Forms]]
+
:* [[Differential Logic and Dynamic Systems 2.0#Tables of Propositional Forms|Tables of Propositional Forms]]
  
==Note 5==
+
==Example 2==
  
For a slightly more interesting example, let's suppose that we have a dynamic system that is known by its state space <math>X,\!</math> and we have a boolean state variable <math>x : X \to \mathbb{B}.</math>  In addition, we are given an initial condition <math>\texttt{x~=~dx}</math> and a law <math>\begin{matrix}\texttt{d}^\texttt{2}\texttt{x~=~(x)}.\end{matrix}</math>
+
For a slightly more interesting example, let's suppose that we have a dynamic system that is known by its state space <math>X\!</math> and we have a boolean state variable <math>x : X \to \mathbb{B}.\!</math>  In addition, we are given an initial condition <math>x = \mathrm{d}x\!</math> and a law <math>{\mathrm{d}^2 x = \texttt{(} x \texttt{)}}.\!</math>
  
 
The initial condition has two cases:
 
The initial condition has two cases:
Line 107: Line 98:
 
|
 
|
 
<math>\begin{array}{ll}
 
<math>\begin{array}{ll}
1. & \texttt{x~=~dx~=~0}
+
1. & x ~=~ \mathrm{d}x ~=~ 0
 
\\
 
\\
2. & \texttt{x~=~dx~=~1}
+
2. & x ~=~ \mathrm{d}x ~=~ 1
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
Here is a table of the two trajectories or ''orbits'' that we get by starting from each of the two permissible initial states and staying within the constraints of the dynamic law <math>\begin{matrix}\texttt{d}^\texttt{2}\texttt{x~=~(x)}.\end{matrix}</math>
+
Here is a table of the two trajectories or ''orbits'' that we get by starting from each of the two permissible initial states and staying within the constraints of the dynamic law <math>{\mathrm{d}^2 x = \texttt{(} x \texttt{)}}.\!</math>
  
 
{| align="center" cellpadding="8" style="text-align:center"
 
{| align="center" cellpadding="8" style="text-align:center"
| <math>\text{Initial State}\ x \cdot dx</math>
+
| <math>\text{Initial State}~ x \cdot \mathrm{d}x\!</math>
 
|-
 
|-
 
|
 
|
<math>\begin{array}{cccc}
+
<math>\begin{array}{c|ccc}
t & d^0 x & d^1 x & d^2 x \\
+
t & \mathrm{d}^0 x & \mathrm{d}^1 x & \mathrm{d}^2 x \\[8pt]
\\
+
0 & 1 & 1 & 0 \\
0 &     1 &     1 &     0 \\
+
1 & 0 & 1 & 1 \\
1 &     0 &     1 &     1 \\
+
2 & 1 & 0 & 0 \\
2 &     1 &     0 &     0 \\
+
3 & 1 & 0 & 0 \\
3 &     1 &     0 &     0 \\
+
4 & 1 & 0 & 0 \\
4 &     1 &     0 &     0 \\
+
5 & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel
5 &     '' &     '' &     '' \\
 
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
Line 134: Line 124:
  
 
{| align="center" cellpadding="8" style="text-align:center"
 
{| align="center" cellpadding="8" style="text-align:center"
| <math>\text{Initial State}\ (x) \cdot (dx)</math>
+
| <math>\text{Initial State}~ \texttt{(} x \texttt{)} \cdot \texttt{(} \mathrm{d}x \texttt{)}\!</math>
 
|-
 
|-
 
|
 
|
<math>\begin{array}{cccc}
+
<math>\begin{array}{c|ccc}
t & d^0 x & d^1 x & d^2 x \\
+
t & \mathrm{d}^0 x & \mathrm{d}^1 x & \mathrm{d}^2 x \\[8pt]
\\
+
0 & 0 & 0 & 1 \\
0 &     0 &     0 &     1 \\
+
1 & 0 & 1 & 1 \\
1 &     0 &     1 &     1 \\
+
2 & 1 & 0 & 0 \\
2 &     1 &     0 &     0 \\
+
3 & 1 & 0 & 0 \\
3 &     1 &     0 &     0 \\
+
4 & 1 & 0 & 0 \\
4 &     1 &     0 &     0 \\
+
5 & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel
5 &     '' &     '' &     '' \\
 
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
Note that the state <math>\begin{matrix}\texttt{x (dx)(d}^\texttt{2}\texttt{x)},\end{matrix}</math> that is, <math>\begin{matrix}(\texttt{x}, \texttt{dx}, \texttt{d}^\texttt{2}\texttt{x}) ~=~ (1, 0, 0),\end{matrix}</math> is a stable attractor for both orbits.
+
Note that the state <math>x \texttt{(} \mathrm{d}x \texttt{)(} \mathrm{d}^2 x \texttt{)},\!</math> that is, <math>(x, \mathrm{d}x, \mathrm{d}^2 x) = (1, 0, 0),\!</math> is a stable attractor for both orbits.
  
 
Further discussion of this example, complete with charts and graphs, can be found at this location:
 
Further discussion of this example, complete with charts and graphs, can be found at this location:
  
:* [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0#Example_1._A_Square_Rigging|Example 1. A Square Rigging]]
+
:* [[Differential Logic and Dynamic Systems 2.0#Example 1. A Square Rigging|Example 1. A Square Rigging]]
  
==Note 6==
+
==Example 3==
  
One more example may serve to suggest just how much dynamic complexity can be built on a universe of discourse that has but a single logical feature at its base.
+
One more example may serve to suggest just how much dynamic complexity can be built on a universe of discourse that has but a single logical feature at its base.  But first, there are a few more elements of general notation that that we'll need to describe finite dimensional universes of discourse and the qualitative dynamics that we envision occurring in them.
  
But first, let me introduce a few more elements of general notation that I'll be using to describe finite dimensional universes of discourse and the qualitative dynamics that we envision occurring in them.
+
Let <math>\mathcal{X} = \{ x_1, \ldots, x_n \}\!</math> be the ''alphabet'' of logical ''features'' or ''variables'' that are used to describe the <math>n\!</math>-dimensional universe of discourse <math>X^\bullet = [\mathcal{X}]\!</math> <math>= [x_1, \ldots, x_n].\!</math>  One may picture a venn diagram whose <math>n\!</math> overlapping &ldquo;circles&rdquo; are labeled with the feature names in <math>\mathcal{X}.\!</math>  Staying with this picture, one visualizes the universe of discourse <math>X^\bullet = [\mathcal{X}]\!</math> as having two layers:
  
Let <math>\mathcal{X} = \{ x_1, \ldots, x_n \}</math> be the ''alphabet'' of logical ''features'' or ''variables'' that we use to describe the n-dimensional universe of discourse <math>X^\circ = [\mathcal{X}] = [ x_1, \ldots, x_n ].</math>  Picturesquely viewed, one may think of a venn diagram with n overlapping "circles" that are labeled with the feature names in the set <math>\mathcal{X}.</math>  Staying with this picture, one visualizes the universe of discourse <math>X^\circ = [\mathcal{X}]</math> as having two layers:  (1) the set <math>X = \langle \mathcal{X} \rangle = \langle x_1, \dots, x_n \rangle</math> of ''points'' or ''cells'' &mdash; in another sense of the word than when we speak of ''cellular automata'' &mdash; (2) the set <math>X^\uparrow = (X \to \mathbb{B})</math> of ''propositions'', boolean-valued functions, or maps from <math>X\!</math> to <math>\mathbb{B}.</math>
+
# The set <math>X = \langle \mathcal{X} \rangle = \langle x_1, \dots, x_n \rangle\!</math> of ''points'' or ''cells'' &mdash; the latter used in another sense of the word than when we speak of ''cellular automata''.
 +
# The set <math>X^\uparrow = (X \to \mathbb{B})\!</math> of ''propositions'', boolean-valued functions, or maps from <math>X\!</math> to <math>\mathbb{B}.\!</math>
  
Thus, we may speak of the universe of discourse <math>X^\circ</math> as being an ordered pair <math>(X, X^\uparrow),</math> with <math>2^n\!</math> points in the underlying space <math>X\!</math> and <math>2^{2^n}</math> propositions in the function space <math>X^\uparrow.</math>
+
Thus we picture the universe of discourse <math>{X^\bullet}\!</math> as an ordered pair <math>{X^\bullet = (X, X^\uparrow)}\!</math> having <math>2^n\!</math> points in the underlying space <math>X\!</math> and <math>2^{2^n}\!</math> propositions in the function space <math>X^\uparrow.\!</math>
  
 
A more complete discussion of these notations can be found here:
 
A more complete discussion of these notations can be found here:
  
:* [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0#A_Functional_Conception_of_Propositional_Calculus|A Functional Conception of Propositional Calculus]]
+
:* [[Differential Logic and Dynamic Systems 2.0#A Functional Conception of Propositional Calculus|A Functional Conception of Propositional Calculus]]
  
 
Now, to the Example.
 
Now, to the Example.
  
Once again, let us begin with a 1-feature alphabet <math>\mathcal{X} = \{ x_1 \} = \{ x \}.</math>  In the discussion that follows I will consider a class of trajectories that are ruled by the constraint that <math>d^k x = 0\!</math> for all <math>k\!</math> greater than some fixed <math>m,\!</math> and I will indulge in the use of some picturesque speech to describes salient classes of such curves.  Given this finite order condition, there is a highest order non-zero difference <math>d^m x\!</math> that is exhibited at each point in the course of any determinate trajectory.  Relative to any point of the corresponding orbit or curve, let us call this highest order differential feature <math>d^m x\!</math> the ''drive'' at that point.  Curves of constant drive <math>d^m x\!</math> are then referred to as ''<math>m^\text{th}\!</math> gear curves''.
+
Once again, let us begin with a 1-feature alphabet <math>\mathcal{X} = \{ x_1 \} = \{ x \}.\!</math>  In the discussion that follows I will consider a class of trajectories that are ruled by the constraint that <math>\mathrm{d}^k x = 0\!</math> for all <math>k\!</math> greater than some fixed <math>m\!</math> and I will indulge in the use of some picturesque language to describe salient classes of such curves.  Given the finite order condition, there is a highest order non-zero difference <math>\mathrm{d}^m x\!</math> that is exhibited at each point in the course of any determinate trajectory.  Relative to any point of the corresponding orbit or curve, let us call this highest order differential feature <math>\mathrm{d}^m x\!</math> the ''drive'' at that point.  Curves of constant drive <math>\mathrm{d}^m x\!</math> are then referred to as ''<math>m^\text{th}\!</math> gear curves''.
 
 
One additional piece of notation will be needed here.  Starting from the base alphabet <math>\mathcal{X} = \{ x \},</math> we define and notate <math>\operatorname{E}^j \mathcal{X} = \{ x, d^1 x, d^2 x, \ldots, d^j x \}</math> as the ''<math>j^\text{th}\!</math> order extended alphabet over <math>\mathcal{X}.</math>''
 
  
Let us now consider the family of <math>4^\text{th}\!</math> gear curves through the extended space <math>\operatorname{E}^4 X = \langle x, dx, d^2 x, d^3 x, d^4 x \rangle.</math> These are the trajectories that are generated subject to the law <math>d^4 x = 1,\!</math> where it is understood in making such a statement that all higher order differences are equal to <math>0.\!</math>
+
One additional piece of notation will be needed here.  Starting from the base alphabet <math>\mathcal{X} = \{ x \},\!</math> we define and notate <math>\mathrm{E}^j \mathcal{X} = \{ x, \mathrm{d}^1 x, \mathrm{d}^2 x, \ldots, \mathrm{d}^j x \}\!</math> as the ''<math>j^\text{th}\!</math> order extended alphabet over <math>\mathcal{X}.\!</math>''
  
Since <math>d^4 x\!</math> and all higher order <math>d^j x\!</math> are fixed, the entire dynamics can be plotted in the extended space <math>\operatorname{E}^3 X = \langle x, dx, d^2 x, d^3 x \rangle.</math>  Thus, there is just enough room in a planar venn diagram to plot both orbits and to show how they partition the points of <math>\operatorname{E}^3 X.</math> As it turns out, there are exactly two possible orbits, of eight points each, as illustrated in Figures&nbsp;16-a and 16-b. See here:
+
Let us now consider the family of 4<sup>th</sup> gear curves through the extended space <math>\mathrm{E}^4 X = \langle x, \mathrm{d}x, \mathrm{d}^2 x, \mathrm{d}^3 x, \mathrm{d}^4 x \rangle.\!</math>  These are the trajectories that are generated subject to the law <math>\mathrm{d}^4 x = 1,\!</math> where it is understood in making such a statement that all higher order differences are equal to <math>0.\!</math>
  
:* [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0#Example_2._Drives_and_Their_Vicissitudes|Example 2. Drives and Their Vicissitudes]]
+
Since <math>\mathrm{d}^4 x\!</math> and all higher order <math>\mathrm{d}^j x\!</math> are fixed, the entire dynamics can be plotted in the extended space <math>\mathrm{E}^3 X = \langle x, \mathrm{d}x, \mathrm{d}^2 x, \mathrm{d}^3 x \rangle.\!</math>  Thus, there is just enough room in a planar venn diagram to plot both orbits and to show how they partition the points of <math>\mathrm{E}^3 X.\!</math>  As it turns out, there are exactly two possible orbits, of eight points each, as illustrated in Figures&nbsp;16-a and 16-b. See here:
  
==Note 7==
+
:* [[Differential Logic and Dynamic Systems 2.0#Example 2. Drives and Their Vicissitudes|Example 2. Drives and Their Vicissitudes]]
  
Here are the <math>4^\text{th}\!</math> gear curves over the 1-feature universe <math>X = \langle x \rangle</math> arranged in the form of tabular arrays, listing the extended state vectors <math>(x, dx, d^2 x, d^3 x, d^4 x)\!</math> as they occur in one cyclic period of each orbit.
+
Here are the 4<sup>th</sup> gear curves over the 1-feature universe <math>X = \langle x \rangle\!</math> arranged in the form of tabular arrays, listing the extended state vectors <math>(x, \mathrm{d}x, \mathrm{d}^2 x, \mathrm{d}^3 x, \mathrm{d}^4 x)\!</math> as they occur in one cyclic period of each orbit.
  
 
{| align="center" cellpadding="8" style="text-align:center"
 
{| align="center" cellpadding="8" style="text-align:center"
Line 190: Line 178:
 
|
 
|
 
<math>\begin{array}{c|ccccc}
 
<math>\begin{array}{c|ccccc}
t & d^0 x & d^1 x & d^2 x & d^3 x & d^4 \\
+
t & \mathrm{d}^0 x & \mathrm{d}^1 x & \mathrm{d}^2 x & \mathrm{d}^3 x & \mathrm{d}^4 x \\
 
\\
 
\\
0 &     0 &     0 &     0 &     0 &   1 \\
+
0 & 0 & 0 & 0 & 0 & 1 \\
1 &     0 &     0 &     0 &     1 &   1 \\
+
1 & 0 & 0 & 0 & 1 & 1 \\
2 &     0 &     0 &     1 &     0 &   1 \\
+
2 & 0 & 0 & 1 & 0 & 1 \\
3 &     0 &     1 &     1 &     1 &   1 \\
+
3 & 0 & 1 & 1 & 1 & 1 \\
4 &     1 &     0 &     0 &     0 &   1 \\
+
4 & 1 & 0 & 0 & 0 & 1 \\
5 &     1 &     0 &     0 &     1 &   1 \\
+
5 & 1 & 0 & 0 & 1 & 1 \\
6 &     1 &     0 &     1 &     0 &   1 \\
+
6 & 1 & 0 & 1 & 0 & 1 \\
7 &     1 &     1 &     1 &     1 &   1 \\
+
7 & 1 & 1 & 1 & 1 & 1
\end{array}</math></p>
+
\end{array}</math>
 
|}
 
|}
  
Line 210: Line 198:
 
|
 
|
 
<math>\begin{array}{c|ccccc}
 
<math>\begin{array}{c|ccccc}
t & d^0 x & d^1 x & d^2 x & d^3 x & d^4 \\
+
t & \mathrm{d}^0 x & \mathrm{d}^1 x & \mathrm{d}^2 x & \mathrm{d}^3 x & \mathrm{d}^4 x \\
 
\\
 
\\
0 &     1 &     1 &     0 &     0 &   1 \\
+
0 & 1 & 1 & 0 & 0 & 1 \\
1 &     0 &     1 &     0 &     1 &   1 \\
+
1 & 0 & 1 & 0 & 1 & 1 \\
2 &     1 &     1 &     1 &     0 &   1 \\
+
2 & 1 & 1 & 1 & 0 & 1 \\
3 &     0 &     0 &     1 &     1 &   1 \\
+
3 & 0 & 0 & 1 & 1 & 1 \\
4 &     0 &     1 &     0 &     0 &   1 \\
+
4 & 0 & 1 & 0 & 0 & 1 \\
5 &     1 &     1 &     0 &     1 &   1 \\
+
5 & 1 & 1 & 0 & 1 & 1 \\
6 &     0 &     1 &     1 &     0 &   1 \\
+
6 & 0 & 1 & 1 & 0 & 1 \\
7 &     1 &     0 &     1 &     1 &   1 \\
+
7 & 1 & 0 & 1 & 1 & 1
\end{array}</math></p>
+
\end{array}</math>
 
|}
 
|}
  
In this arrangement, the temporal ordering of states can be reckoned by a kind of ''parallel round-up rule''.  Specifically, if <math>(a_k, a_{k+1})\!</math> is any pair of adjacent digits in a state vector <math>(a_0, a_1, \ldots, a_n),\!</math> then the value of <math>a_k\!</math> in the next state is <math>a_k^\prime = a_k + a_{k+1},\!</math> the addition being taken mod 2, of course.
+
In this arrangement, the temporal ordering of states can be reckoned by a kind of ''parallel round-up rule''.  Specifically, if <math>(a_k, a_{k+1})\!</math> is any pair of adjacent digits in a state vector <math>{(a_0, a_1, \ldots, a_n)},\!</math> then the value of <math>a_k\!</math> in the next state is <math>a_k' = a_k + a_{k+1},\!</math> the addition being taken mod 2, of course.
  
 
A more complete discussion of this arrangement is given here:
 
A more complete discussion of this arrangement is given here:
  
:* [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0#Example_2._Drives_and_Their_Vicissitudes|Example 2. Drives and Their Vicissitudes]]
+
:* [[Differential Logic and Dynamic Systems 2.0#Example 2. Drives and Their Vicissitudes|Example 2. Drives and Their Vicissitudes]]
  
==Note 8==
+
==Example 4==
  
 
I am going to tip-toe in silence/consilience past many questions of a philosophical nature/nurture that might be asked at this juncture, no doubt to revisit them at some future opportunity/importunity, however the cases happen to align in the course of their inevitable fall.
 
I am going to tip-toe in silence/consilience past many questions of a philosophical nature/nurture that might be asked at this juncture, no doubt to revisit them at some future opportunity/importunity, however the cases happen to align in the course of their inevitable fall.
  
Instead, let's "keep it concrete and simple", taking up the consideration of an incrementally more complex example, but having a slightly more general character than the orders of sequential transformations that we've been discussing up to this point.
+
Instead, let's follow the adage to &ldquo;keep it concrete and simple&rdquo;, taking up the consideration of an incrementally more complex example, but having a slightly more general character than the orders of sequential transformations that we've been discussing up to this point.
  
 
The types of logical transformations that I have in mind can be thought of as ''transformations of discourse'' because they map a universe of discourse into a universe of discourse by way of logical equations between the qualitative features or logical variables in the source and target universes.
 
The types of logical transformations that I have in mind can be thought of as ''transformations of discourse'' because they map a universe of discourse into a universe of discourse by way of logical equations between the qualitative features or logical variables in the source and target universes.
Line 241: Line 229:
 
Onward and upward to Flatland, the differential analysis of transformations between 2-dimensional universes of discourse.
 
Onward and upward to Flatland, the differential analysis of transformations between 2-dimensional universes of discourse.
  
Consider the transformation from the universe <math>U^\circ = [u, v]</math> to the universe <math>X^\circ = [x, y]</math> that is defined by this system of equations:
+
Consider the transformation from the universe <math>U^\bullet = [u, v]</math> to the universe <math>X^\bullet = [x, y]</math> that is defined by this system of equations:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{array}{lcccc}
+
<math>\begin{matrix}
x & = & f(u, v) & = & \texttt{((u)(v))}
+
x & = & f(u, v) & = & \texttt{((} u \texttt{)(} v \texttt{))}
\\ \\
+
\\[8pt]
y & = & g(u, v) & = & \texttt{((u,~v))}
+
y & = & g(u, v) & = & \texttt{((} u \texttt{,~} v \texttt{))}
\end{array}</math>
+
\end{matrix}</math>
 
|}
 
|}
  
The underlined parenthetical expressions on the right are the cactus forms for the boolean functions that correspond to inclusive disjunction and logical equivalence, respectively.  Table&nbsp;1 summarizes the basic elements of the cactus notation for propositional logic.
+
The parenthetical expressions on the right are the cactus forms for the boolean functions that correspond to inclusive disjunction and logical equivalence, respectively.  Table&nbsp;1 summarizes the basic elements of the cactus notation for propositional logic.
  
 
<br>
 
<br>
  
{| align="center" border="1" cellpadding="8" cellspacing="0" style="text-align:center; width:96%"
+
{| align="center" border="1" cellpadding="8" cellspacing="0" style="text-align:center; width:75%"
|+ <math>\text{Table 1. Syntax and Semantics of a Calculus for Propositional Logic}\!</math>
+
|+ style="height:30px" | <math>\text{Table 1.} ~~ \text{Syntax and Semantics of a Calculus for Propositional Logic}\!</math>
|- style="background:whitesmoke"
+
|- style="height:40px; background:ghostwhite"
| <math>\text{Expression}\!</math>
+
| <math>\text{Graph}\!</math>
 +
| <math>\text{Expression}~\!</math>
 
| <math>\text{Interpretation}\!</math>
 
| <math>\text{Interpretation}\!</math>
 
| <math>\text{Other Notations}\!</math>
 
| <math>\text{Other Notations}\!</math>
 
|-
 
|-
| <math>\texttt{~}</math>
+
| height="100px" | [[Image:Cactus Node Big Fat.jpg|20px]]
| <math>\operatorname{True}</math>
+
| <math>~</math>
 +
| <math>\operatorname{true}</math>
 
| <math>1\!</math>
 
| <math>1\!</math>
 
|-
 
|-
 +
| height="100px" | [[Image:Cactus Spike Big Fat.jpg|20px]]
 
| <math>\texttt{(~)}</math>
 
| <math>\texttt{(~)}</math>
| <math>\operatorname{False}</math>
+
| <math>\operatorname{false}</math>
 
| <math>0\!</math>
 
| <math>0\!</math>
 
|-
 
|-
| <math>\texttt{x}</math>
+
| height="100px" | [[Image:Cactus A Big.jpg|20px]]
| <math>x\!</math>
+
| <math>a\!</math>
| <math>x\!</math>
+
| <math>a\!</math>
 +
| <math>a\!</math>
 
|-
 
|-
| <math>\texttt{(x)}</math>
+
| height="120px" | [[Image:Cactus (A) Big.jpg|20px]]
| <math>\operatorname{Not}\ x</math>
+
| <math>\texttt{(} a \texttt{)}~</math>
|
+
| <math>\operatorname{not}~ a</math>
<math>\begin{matrix}
+
| <math>\lnot a \quad \bar{a} \quad \tilde{a} \quad a^\prime</math>
x'        \\
 
\tilde{x} \\
 
\lnot x  \\
 
\end{matrix}</math>
 
 
|-
 
|-
| <math>\texttt{x~y~z}</math>
+
| height="100px" | [[Image:Cactus ABC Big.jpg|50px]]
| <math>x\ \operatorname{and}\ y\ \operatorname{and}\ z</math>
+
| <math>a ~ b ~ c</math>
| <math>x \land y \land z</math>
+
| <math>a ~\operatorname{and}~ b ~\operatorname{and}~ c</math>
 +
| <math>a \land b \land c</math>
 
|-
 
|-
| <math>\texttt{((x)(y)(z))}</math>
+
| height="160px" | [[Image:Cactus ((A)(B)(C)) Big.jpg|65px]]
| <math>x\ \operatorname{or}\ y\ \operatorname{or}\ z</math>
+
| <math>\texttt{((} a \texttt{)(} b \texttt{)(} c \texttt{))}</math>
| <math>x \lor y \lor z</math>
+
| <math>a ~\operatorname{or}~ b ~\operatorname{or}~ c</math>
 +
| <math>a \lor b \lor c</math>
 
|-
 
|-
| <math>\texttt{(x~(y))}</math>
+
| height="120px" | [[Image:Cactus (A(B)) Big.jpg|60px]]
 +
| <math>\texttt{(} a \texttt{(} b \texttt{))}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
x\ \operatorname{implies}\ y                \\
+
a ~\operatorname{implies}~ b
\operatorname{If}\ x\ \operatorname{then}\ y \\
+
\\[6pt]
 +
\operatorname{if}~ a ~\operatorname{then}~ b
 
\end{matrix}</math>
 
\end{matrix}</math>
| <math>x \Rightarrow y\!</math>
+
| <math>a \Rightarrow b</math>
 
|-
 
|-
| <math>\texttt{(x,~y)}</math>
+
| height="120px" | [[Image:Cactus (A,B) Big ISW.jpg|65px]]
 +
| <math>\texttt{(} a \texttt{,} b \texttt{)}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
x\ \operatorname{not~equal~to}\ y \\
+
a ~\operatorname{not~equal~to}~ b
x\ \operatorname{exclusive~or}\ y \\
+
\\[6pt]
 +
a ~\operatorname{exclusive~or}~ b
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
x \neq y \\
+
a \neq b
x + y    \\
+
\\[6pt]
 +
a + b
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|-
 
|-
| <math>\texttt{((x,~y))}</math>
+
| height="160px" | [[Image:Cactus ((A,B)) Big.jpg|65px]]
 +
| <math>\texttt{((} a \texttt{,} b \texttt{))}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
x\ \operatorname{is~equal~to}\ y    \\
+
a ~\operatorname{is~equal~to}~ b
x\ \operatorname{if~and~only~if}\ y \\
+
\\[6pt]
 +
a ~\operatorname{if~and~only~if}~ b
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
x = y              \\
+
a = b
x \Leftrightarrow y \\
+
\\[6pt]
 +
a \Leftrightarrow b
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|-
 
|-
| <math>\texttt{(x,~y,~z)}</math>
+
| height="120px" | [[Image:Cactus (A,B,C) Big.jpg|65px]]
 +
| <math>\texttt{(} a \texttt{,} b \texttt{,} c \texttt{)}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
\operatorname{Just~one~of} \\
+
\operatorname{just~one~of}
x, y, z                    \\
+
\\
\operatorname{is~false}.   \\
+
a, b, c
 +
\\
 +
\operatorname{is~false}.
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
x'y~z~ & \lor \\
+
& \bar{a} ~ b ~ c
x~y'z~ & \lor \\
+
\\
x~y~z' &      \\
+
\lor & a ~ \bar{b} ~ c
 +
\\
 +
\lor & a ~ b ~ \bar{c}
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|-
 
|-
| <math>\texttt{((x),(y),(z))}</math>
+
| height="160px" | [[Image:Cactus ((A),(B),(C)) Big.jpg|65px]]
 +
| <math>\texttt{((} a \texttt{),(} b \texttt{),(} c \texttt{))}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
\operatorname{Just~one~of}   \\
+
\operatorname{just~one~of}
x, y, z                      \\
+
\\
\operatorname{is~true}.       \\
+
a, b, c
&                            \\
+
\\
\operatorname{Partition~all} \\
+
\operatorname{is~true}.
\operatorname{into}\ x, y, z. \\
+
\\[6pt]
 +
\operatorname{partition~all}
 +
\\
 +
\operatorname{into}~ a, b, c.
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
x~y'z' & \lor \\
+
& a ~ \bar{b} ~ \bar{c}
x'y~z' & \lor \\
+
\\
x'y'z~ &     \\
+
\lor & \bar{a} ~ b ~ \bar{c}
 +
\\
 +
\lor & \bar{a} ~ \bar{b} ~ c
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|-
 
|-
 +
| height="160px" | [[Image:Cactus (A,(B,C)) Big.jpg|90px]]
 +
| <math>\texttt{(} a \texttt{,(} b \texttt{,} c \texttt{))}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
\texttt{((x,~y),~z)}
+
\operatorname{oddly~many~of}
\\ \\
+
\\
\texttt{(x,~(y,~z))}
+
a, b, c
\end{matrix}</math>
+
\\
|
+
\operatorname{are~true}.
<math>\begin{matrix}
 
\operatorname{Oddly~many~of} \\
 
x, y, z                      \\
 
\operatorname{are~true}.     \\
 
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|
 
|
<p><math>x + y + z\!</math></p>
+
<p><math>a + b + c\!</math></p>
 
<br>
 
<br>
 
<p><math>\begin{matrix}
 
<p><math>\begin{matrix}
x~y~z~ & \lor \\
+
& a ~ b ~ c
x~y'z' & \lor \\
+
\\
x'y~z' & \lor \\
+
\lor & a ~ \bar{b} ~ \bar{c}
x'y'z~ &     \\
+
\\
 +
\lor & \bar{a} ~ b ~ \bar{c}
 +
\\
 +
\lor & \bar{a} ~ \bar{b} ~ c
 
\end{matrix}</math></p>
 
\end{matrix}</math></p>
 
|-
 
|-
| <math>\texttt{(w,~(x),(y),(z))}</math>
+
| height="160px" | [[Image:Cactus (X,(A),(B),(C)) Big.jpg|90px]]
 +
| <math>\texttt{(} x \texttt{,(} a \texttt{),(} b \texttt{),(} c \texttt{))}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
\operatorname{Partition}\ w      \\
+
\operatorname{partition}~ x
\operatorname{into}\ x, y, z.   \\
+
\\
&                                \\
+
\operatorname{into}~ a, b, c.
\operatorname{Genus}\ w\ \operatorname{comprises} \\
+
\\[6pt]
\operatorname{species}\ x, y, z. \\
+
\operatorname{genus}~ x ~\operatorname{comprises}
 +
\\
 +
\operatorname{species}~ a, b, c.
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
w'x'y'z' & \lor \\
+
& \bar{x} ~ \bar{a} ~ \bar{b} ~ \bar{c}
w~x~y'z' & \lor \\
+
\\
w~x'y~z' & \lor \\
+
\lor & x ~ a ~ \bar{b} ~ \bar{c}
w~x'y'z~ &      \\
+
\\
 +
\lor & x ~ \bar{a} ~ b ~ \bar{c}
 +
\\
 +
\lor & x ~ \bar{a} ~ \bar{b} ~ c
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|}
 
|}
Line 397: Line 413:
 
<br>
 
<br>
  
The component notation <math>F = (F_1, F_2) = (f, g) : U^\circ \to X^\circ</math> allows us to give a name and a type to this transformation, and permits us to define it by means of the compact description that follows:
+
The component notation <math>F = (F_1, F_2) = (f, g) : U^\bullet \to X^\bullet\!</math> allows us to give a name and a type to this transformation, and permits us to define it by means of the compact description that follows:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{array}{lcccc}
+
<math>\begin{matrix}
(x, y) & = & F(u, v) & = & ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ ).
+
(x, y) & = & F(u, v) & = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ ).
\end{array}</math>
+
\end{matrix}</math>
 
|}
 
|}
  
The information that defines the logical transformation <math>F\!</math> can be represented in the form of a truth table, as below.
+
The information that defines the logical transformation <math>F\!</math> can be represented in the form of a truth table, as shown below.
  
 
<br>
 
<br>
  
{| align="center" cellpadding="8" cellspacing="0" style="border-left:1px solid black; border-top:1px solid black; border-right:1px solid black; border-bottom:1px solid black; text-align:center; width:40%"
+
{| align="center" cellpadding="8" cellspacing="0" style="border-bottom:1px solid black; border-left:1px solid black; border-right:1px solid black; border-top:1px solid black; text-align:center; width:60%"
 +
|- style="height:40px; background:ghostwhite; width:100%"
 +
| style="width:25%" | <math>u\!</math>
 +
| style="width:25%" | <math>v\!</math>
 +
| style="width:25%; border-left:1px solid black" | <math>f\!</math>
 +
| style="width:25%" | <math>g\!</math>
 
|-
 
|-
| style="border-bottom:1px solid black" | <math>u\!</math>
+
| style="border-top:1px solid black" |
| style="border-bottom:1px solid black" | <math>v\!</math>
+
<math>\begin{matrix}
| style="border-bottom:1px solid black; border-left:1px solid black" | <math>f\!</math>
+
0
| style="border-bottom:1px solid black" | <math>g\!</math>
+
\\[4pt]
|-
+
0
| <math>0\!</math>
+
\\[4pt]
| <math>0\!</math>
+
1
| style="border-left:1px solid black" | <math>0\!</math>
+
\\[4pt]
| <math>1\!</math>
+
1
|-
+
\end{matrix}</math>
| <math>0\!</math>
+
| style="border-top:1px solid black" |
| <math>1\!</math>
+
<math>\begin{matrix}
| style="border-left:1px solid black" | <math>1\!</math>
+
0
| <math>0\!</math>
+
\\[4pt]
|-
+
1
| <math>1\!</math>
+
\\[4pt]
| <math>0\!</math>
+
0
| style="border-left:1px solid black" | <math>1\!</math>
+
\\[4pt]
| <math>0\!</math>
+
1
|-
+
\end{matrix}</math>
| <math>1\!</math>
+
| style="border-top:1px solid black; border-left:1px solid black" |
| <math>1\!</math>
+
<math>\begin{matrix}
| style="border-left:1px solid black" | <math>1\!</math>
+
0
| <math>1\!</math>
+
\\[4pt]
 +
1
 +
\\[4pt]
 +
1
 +
\\[4pt]
 +
1
 +
\end{matrix}</math>
 +
| style="border-top:1px solid black" |
 +
<math>\begin{matrix}
 +
1
 +
\\[4pt]
 +
0
 +
\\[4pt]
 +
0
 +
\\[4pt]
 +
1
 +
\end{matrix}</math>
 +
|- style="height:40px; background:ghostwhite"
 +
| style="border-top:1px solid black" | <math>u\!</math>
 +
| style="border-top:1px solid black" | <math>v\!</math>
 +
| style="border-top:1px solid black; border-left:1px solid black" |
 +
<math>\texttt{((} u \texttt{)(} v \texttt{))}\!</math>
 +
| style="border-top:1px solid black" |
 +
<math>\texttt{((} u \texttt{,} v \texttt{))}\!</math>
 
|}
 
|}
  
Line 442: Line 486:
 
A more complete framework of discussion and a fuller development of this example can be found in the neighborhood of the following site:
 
A more complete framework of discussion and a fuller development of this example can be found in the neighborhood of the following site:
  
:* [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0#Transformations_of_Type_B2_.E2.86.92_B2|Transformations of Type '''B'''<sup>2</sup> &rarr; '''B'''<sup>2</sup>]]
+
:* [[Differential Logic and Dynamic Systems 2.0#Transformations of Type B2 .E2.86.92 B2|Transformations of Type '''B'''<sup>2</sup> &rarr; '''B'''<sup>2</sup>]]
 
 
==Note 9==
 
  
Consider the "transformation of textual elements" (TOTE) in progress:
+
Consider the ''transformation of textual elements'' (TOTE) in progress:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{array}{ccccc}
+
<math>\begin{matrix}
x     & = & f(u, v) & = & \texttt{((u)(v))}
+
x
\\ \\
+
& = & f(u, v)
y     & = & g(u, v) & = & \texttt{((u,~v))}
+
& = & \texttt{((} u \texttt{)(} v \texttt{))}
\\ \\
+
\\[8pt]
(x, y) & = & F(u, v) & = & ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ )
+
y
\end{array}</math>
+
& = & g(u, v)
 +
& = & \texttt{((} u \texttt{,~} v \texttt{))}
 +
\\[8pt]
 +
(x, y)
 +
& = & F(u, v)
 +
& = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ )
 +
\end{matrix}</math>
 
|}
 
|}
  
Taken as a transformation from the universe <math>U^\circ = [u, v]</math> to the universe <math>X^\circ = [x, y],</math> this is a particular type of formal object, and it can be studied at that level of abstraction until the chickens come home to roost, as they say, but when the time comes to count those chickens, if you will, the terms of artifice that we use to talk about abstract objects, almost as if we actually knew what we were talking about, need to be fully fledged or fleshed out with extra "bits of interpretive data" (BOIDs).
+
Taken as a transformation from the universe <math>U^\bullet = [u, v]\!</math> to the universe <math>X^\bullet = [x, y],\!</math> this is a particular type of formal object, and it can be studied at that level of abstraction until the chickens come home to roost, as they say, but when the time comes to count those chickens, if you will, the terms of artifice that we use to talk about abstract objects, almost as if we actually knew what we were talking about, need to be fully fledged or fleshed out with extra ''bits of interpretive data'' (BOIDs).
  
 
And so, to decompress the story, the TOTE that we use to convey the FOMA has to be interpreted before it can be applied to anything that actually puts supper on the table, so to speak.
 
And so, to decompress the story, the TOTE that we use to convey the FOMA has to be interpreted before it can be applied to anything that actually puts supper on the table, so to speak.
Line 469: Line 517:
 
When we consider a transformation in the alias interpretation, we are merely changing the terms that we use to describe what may very well be, to some approximation, the very same things.
 
When we consider a transformation in the alias interpretation, we are merely changing the terms that we use to describe what may very well be, to some approximation, the very same things.
  
For example, in some applications the discursive universes <math>U^\circ = [u, v]</math> and <math>X^\circ = [x, y]</math> are best understood as diverse frames, instruments, reticules, scopes, or templates, that we adopt for the sake of viewing from variant perspectives what we conceive to be roughly the same underlying objects.
+
For example, in some applications the discursive universes <math>U^\bullet = [u, v]\!</math> and <math>X^\bullet = [x, y]\!</math> are best understood as diverse frames, instruments, reticules, scopes, or templates, that we adopt for the sake of viewing from variant perspectives what we conceive to be roughly the same underlying objects.
  
 
When we consider a transformation in the alibi interpretation, we are thinking of the objective things as objectively moving around in space or changing their qualitative characteristics.  There are times when we think of this alibi transformation as taking place in a dimension of time, and then there are times when time is not an object.
 
When we consider a transformation in the alibi interpretation, we are thinking of the objective things as objectively moving around in space or changing their qualitative characteristics.  There are times when we think of this alibi transformation as taking place in a dimension of time, and then there are times when time is not an object.
  
For example, in some applications the discursive universes <math>U^\circ = [u, v]</math> and <math>X^\circ = [x, y]</math> are actually the same universe, and what we have is a frame where <math>x\!</math> is the next state of <math>u\!</math> and <math>y\!</math> is the next state of <math>v,\!</math> notated as <math>x = u'\!</math> and <math>y = v'.\!</math>  This permits us to rewrite the transformation <math>F\!</math> as follows:
+
For example, in some applications the discursive universes <math>U^\bullet = [u, v]\!</math> and <math>X^\bullet = [x, y]\!</math> are actually the same universe, and what we have is a frame where <math>x\!</math> is the next state of <math>u\!</math> and <math>y\!</math> is the next state of <math>v,\!</math> notated as <math>x = u'\!</math> and <math>y = v'.\!</math>  This permits us to rewrite the transformation <math>F\!</math> as follows:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{array}{ccccc}
+
<math>\begin{matrix}
u'       & = & f(u, v) & = & \texttt{((u)(v))}
+
u'
\\ \\
+
& = & f(u, v)
v'       & = & g(u, v) & = & \texttt{((u,~v))}
+
& = & \texttt{((} u \texttt{)(} v \texttt{))}
\\ \\
+
\\[8pt]
(u', v') & = & F(u, v) & = & ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ )
+
v'
\end{array}</math>
+
& = & g(u, v)
 +
& = & \texttt{((} u \texttt{,~} v \texttt{))}
 +
\\[8pt]
 +
(u', v')
 +
& = & F(u, v)
 +
& = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ )
 +
\end{matrix}</math>
 
|}
 
|}
  
All in all, then, we have three different ways in general of applying or interpreting a transformation of discourse, that we might sum up as one brand of alias and two brands
+
All in all, then, we have three different ways in general of applying or interpreting a transformation of discourse, that we might sum up as one brand of alias and two brands of alibi, all together, the ''Elseword'', the ''Elsewhere'', and the ''Elsewhen''.
of alibi, all together, the ''Elseword'', the ''Elsewhere'', and the ''Elsewhen''.
 
  
 
No more angels on pinheads, the brass tacks next time.
 
No more angels on pinheads, the brass tacks next time.
  
==Note 10==
+
==Differential Analysis==
  
 
It is time to formulate the differential analysis of a logical transformation, or a ''mapping of discourse''.  It is wise to begin with the first order differentials.
 
It is time to formulate the differential analysis of a logical transformation, or a ''mapping of discourse''.  It is wise to begin with the first order differentials.
  
We are considering an abstract logical transformation <math>F = (f, g) : [u, v] \to [x, y]</math> that can be interpreted in a number of different ways.  Let's fix on a couple of major variants that might be indicated as follows:
+
We are considering an abstract logical transformation <math>F = (f, g) : [u, v] \to [x, y]\!</math> that can be interpreted in a number of different ways.  Let's fix on a couple of major variants that might be indicated as follows:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
 
<math>\begin{array}{lccccc}
 
<math>\begin{array}{lccccc}
\text{Alias Map.} & (x, y)   & = & F(u, v) & = & ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ )
+
\text{Alias Map}
\\ \\
+
& (x, y)
\text{Alibi Map.} & (u', v') & = & F(u, v) & = & ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ )
+
& = & F(u, v)
 +
& = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ )
 +
\\[8pt]
 +
\text{Alibi Map}
 +
& (u', v')
 +
& = & F(u, v)
 +
& = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ )
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
Line 508: Line 567:
 
<math>F\!</math> is just one example among &mdash; well, now that I think of it &mdash; how many other logical transformations from the same source to the same target universe?  In the light of that question, maybe it would be advisable to contemplate the character of <math>F\!</math> within the fold of its most closely akin transformations.
 
<math>F\!</math> is just one example among &mdash; well, now that I think of it &mdash; how many other logical transformations from the same source to the same target universe?  In the light of that question, maybe it would be advisable to contemplate the character of <math>F\!</math> within the fold of its most closely akin transformations.
  
Given the alphabets <math>\mathcal{U} = \{ u, v \}</math> and <math>\mathcal{X} = \{ x, y \},</math> along with the corresponding universes of discourse <math>U^\circ</math> and <math>X^\circ = [\mathbb{B}^2],</math> how many logical transformations of the general form <math>G = (G_1, G_2) : U^\circ \to X^\circ</math> are there?
+
Given the alphabets <math>\mathcal{U} = \{ u, v \}\!</math> and <math>\mathcal{X} = \{ x, y \}\!</math> along with the corresponding universes of discourse <math>U^\bullet\!</math> and <math>X^\bullet = [\mathbb{B}^2],\!</math> how many logical transformations of the general form <math>G = (G_1, G_2) : U^\bullet \to X^\bullet\!</math> are there?
  
Since <math>G_1\!</math> and <math>G_2\!</math> can be any propositions of the type <math>\mathbb{B}^2 \to \mathbb{B},</math> there are <math>2^4 = 16\!</math> choices for each of the maps <math>G_1\!</math> and <math>G_2,\!</math> and thus there are <math>2^4 \cdot 2^4 = 2^8 = 256\!</math> different mappings altogether of the form <math>G : U^\circ \to X^\circ.</math>
+
Since <math>G_1\!</math> and <math>G_2\!</math> can be any propositions of the type <math>\mathbb{B}^2 \to \mathbb{B},\!</math> there are <math>2^4 = 16\!</math> choices for each of the maps <math>G_1\!</math> and <math>G_2,\!</math> and thus there are <math>2^4 \cdot 2^4 = 2^8 = 256\!</math> different mappings altogether of the form <math>G : U^\bullet \to X^\bullet.\!</math>
  
The set of all functions of a given type is customarily denoted by placing its type indicator in parentheses, in the present instance writing <math>(U^\circ \to X^\circ) = \{ G : U^\circ \to X^\circ \},</math> and so the cardinality of this ''function space'' can most conveniently be summed up by writing:
+
The set of all functions of a given type is customarily denoted by placing its type indicator in parentheses, in the present instance writing <math>(U^\bullet \to X^\bullet) = \{ G : U^\bullet \to X^\bullet \},\!</math> and so the cardinality of this ''function space'' can most conveniently be summed up by writing:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
| <math>|(U^\circ \to X^\circ)| ~=~ |(\mathbb{B}^2 \to \mathbb{B}^2)| ~=~ 4^4 ~=~ 256.</math>
+
| <math>|(U^\bullet \to X^\bullet)| ~=~ |(\mathbb{B}^2 \to \mathbb{B}^2)| ~=~ 4^4 ~=~ 256.\!</math>
 
|}
 
|}
  
Given any transformation of this type, <math>G : U^\circ \to X^\circ,</math> the (first order)
+
Given any transformation of this type, <math>G : U^\bullet \to X^\bullet,\!</math> the (first order) differential analysis of <math>G\!</math> is based on the definition of a couple of further transformations, derived by way of operators on <math>G,\!</math> that ply between the (first order) extended universes, <math>\mathrm{E}U^\bullet = [u, v, du, dv]\!</math> and <math>\mathrm{E}X^\bullet = [x, y, dx, dy],\!</math> of <math>G\text{'s}\!</math> own source and target universes.
differential analysis of <math>G\!</math> is based on the definition of a couple of further transformations, derived by way of operators on <math>G,\!</math> that ply between the (first order) extended universes, <math>\operatorname{E}U^\circ = [u, v, du, dv]</math> and <math>\operatorname{E}X^\circ = [x, y, dx, dy],</math> of <math>G \operatorname{'s}\!</math> own source and target universes.
 
  
First, the ''enlargement map'' (or the ''secant transformation'') <math>\operatorname{E}G = (\operatorname{E}G_1, \operatorname{E}G_2) : \operatorname{E}U^\circ \to \operatorname{E}X^\circ</math> is defined by the following pair of component equations:
+
First, the ''enlargement map'' (or the ''secant transformation'') <math>\mathrm{E}G = (\mathrm{E}G_1, \mathrm{E}G_2) : \mathrm{E}U^\bullet \to \mathrm{E}X^\bullet\!</math> is defined by the following pair of component equations:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
 
<math>\begin{array}{lll}
 
<math>\begin{array}{lll}
\operatorname{E}G_1 & = & G_1 (u + du, v + dv)
+
\mathrm{E}G_1
\\ \\
+
& = & G_1 (u + \mathrm{d}u, v + \mathrm{d}v)
\operatorname{E}G_2 & = & G_2 (u + du, v + dv)
+
\\[8pt]
 +
\mathrm{E}G_2
 +
& = & G_2 (u + \mathrm{d}u, v + \mathrm{d}v)
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
Second, the ''difference map'' (or the ''chordal transformation'') <math>\operatorname{D}G = (\operatorname{D}G_1, \operatorname{D}G_2) : \operatorname{E}U^\circ \to \operatorname{E}X^\circ</math> is defined in a component-wise fashion as the boolean sum of the initial proposition <math>G_j\!</math> and the ''enlarged'' or ''shifted'' proposition <math>\operatorname{E}G_j,</math> for <math>j = 1, 2,\!</math> in accord with following pair of equations:
+
Second, the ''difference map'' (or the ''chordal transformation'') <math>{\mathrm{D}G = (\mathrm{D}G_1, \mathrm{D}G_2) : \mathrm{E}U^\bullet \to \mathrm{E}X^\bullet}\!</math> is defined in a component-wise fashion as the boolean sum of the initial proposition <math>G_j\!</math> and the ''enlarged'' or ''shifted'' proposition <math>\mathrm{E}G_j,\!</math> for <math>j = 1, 2,\!</math> in accord with following pair of equations:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
 
<math>\begin{array}{lllll}
 
<math>\begin{array}{lllll}
\operatorname{D}G_1 & = & G_1 (u, v) & + & \operatorname{E}G_1 (u, v, du, dv)
+
\mathrm{D}G_1
\\ \\
+
& = & G_1 (u, v)
                    & = & G_1 (u, v) & + & G_1 (u + du, v + dv)
+
& + & \mathrm{E}G_1 (u, v, \mathrm{d}u, \mathrm{d}v)
\\ \\
+
\\[8pt]
\operatorname{D}G_2 & = & G_2 (u, v) & + & \operatorname{E}G_2 (u, v, du, dv)
+
& = & G_1 (u, v)
\\ \\
+
& + & G_1 (u + \mathrm{d}u, v + \mathrm{d}v)
                    & = & G_2 (u, v) & + & G_2 (u + du, v + dv)
+
\\[8pt]
 +
\mathrm{D}G_2
 +
& = & G_2 (u, v)
 +
& + & \mathrm{E}G_2 (u, v, \mathrm{d}u, \mathrm{d}v)
 +
\\[8pt]
 +
& = & G_2 (u, v)
 +
& + & G_2 (u + \mathrm{d}u, v + \mathrm{d}v)
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
Maintaining a strict analogy with ordinary difference calculus would perhaps have us write <math>\operatorname{D}G_j = \operatorname{E}G_j - G_j,</math> but the sum and difference operations are the same thing in boolean arithmetic.  It is more often natural in the logical context to consider an initial proposition <math>q,\!</math> then to compute the enlargement <math>\operatorname{E}q,</math> and finally to determine the difference <math>\operatorname{D}q = q + \operatorname{E}q,</math> so we let the variant order of terms reflect this sequence of considerations.
+
Maintaining a strict analogy with ordinary difference calculus would perhaps have us write <math>\mathrm{D}G_j = \mathrm{E}G_j - G_j,</math> but the sum and difference operations are the same thing in boolean arithmetic.  It is more often natural in the logical context to consider an initial proposition <math>q,\!</math> then to compute the enlargement <math>\mathrm{E}q,</math> and finally to determine the difference <math>\mathrm{D}q = q + \mathrm{E}q,\!</math> so we let the variant order of terms reflect this sequence of considerations.
 
 
Given these general considerations about the operators <math>\operatorname{E}</math> and <math>\operatorname{D},</math> let's return to particular cases, and carry out the first order analysis of the transformation <math>F(u, v) ~=~ ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ ).</math>
 
  
==Note 11==
+
Given these general considerations about the operators <math>\mathrm{E}\!</math> and <math>\mathrm{D},\!</math> let's return to particular cases, and carry out the first order analysis of the transformation <math>F(u, v) = ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ ).\!</math>
  
By way of getting our feet back on solid ground, let's crank up our current case of a transformation of discourse, <math>F : U^\circ \to X^\circ,</math> with concrete type <math>[u, v] \to [x, y]</math> or abstract type <math>\mathbb{B}^2 \to \mathbb{B}^2,</math> and let it spin through a sufficient number of turns to see how it goes, as viewed under the scope of what is probably its most straightforward view, as an elsewhen map <math>F : [u, v] \to [u', v'].</math>
+
By way of getting our feet back on solid ground, let's crank up our current case of a transformation of discourse, <math>F : U^\bullet \to X^\bullet,\!</math> with concrete type <math>[u, v] \to [x, y]\!</math> or abstract type <math>\mathbb{B}^2 \to \mathbb{B}^2,\!</math> and let it spin through a sufficient number of turns to see how it goes, as viewed under the scope of what is probably its most straightforward view, as an elsewhen map <math>F : [u, v] \to [u', v'].\!</math>
  
 
{| align="center" cellpadding="8" style="text-align:center"
 
{| align="center" cellpadding="8" style="text-align:center"
 
|
 
|
 
<math>\begin{array}{ccc}
 
<math>\begin{array}{ccc}
\texttt{u}' & = & \texttt{((u)(v))}
+
u' & = & \texttt{((} u \texttt{)(} v \texttt{))}
 
\\
 
\\
\texttt{v}' & = & \texttt{((u,~v))}
+
v' & = & \texttt{((} u \texttt{,~} v \texttt{))}
 
\end{array}</math>
 
\end{array}</math>
 
|-
 
|-
Line 572: Line 636:
 
|
 
|
 
<math>\begin{array}{c|cc}
 
<math>\begin{array}{c|cc}
t & u & v \\
+
t & u & v \\[8pt]
\\
+
0 & 1 & 1 \\
0 & 1 & 1 \\
+
1 & 1 & 1 \\
1 & 1 & 1 \\
+
2 & {}^\shortparallel & {}^\shortparallel
2 & '' & '' \\
 
 
\end{array}</math>
 
\end{array}</math>
 
|-
 
|-
Line 588: Line 651:
 
|
 
|
 
<math>\begin{array}{c|cc}
 
<math>\begin{array}{c|cc}
t & u & v \\
+
t & u & v \\[8pt]
\\
+
0 & 0 & 0 \\
0 & 0 & 0 \\
+
1 & 0 & 1 \\
1 & 0 & 1 \\
+
2 & 1 & 0 \\
2 & 1 & 0 \\
+
3 & 1 & 0 \\
3 & 1 & 0 \\
+
4 & {}^\shortparallel & {}^\shortparallel
4 & '' & '' \\
 
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
Line 600: Line 662:
 
In the upshot there are two basins of attraction, the state <math>(1, 1)\!</math> and the state <math>(1, 0),\!</math> with the orbit <math>(1, 1)\!</math> making up an isolated basin and the orbit <math>(0, 0), (0, 1), (1, 0)\!</math> leading to the basin <math>(1, 0).\!</math>
 
In the upshot there are two basins of attraction, the state <math>(1, 1)\!</math> and the state <math>(1, 0),\!</math> with the orbit <math>(1, 1)\!</math> making up an isolated basin and the orbit <math>(0, 0), (0, 1), (1, 0)\!</math> leading to the basin <math>(1, 0).\!</math>
  
==Note 12==
+
On first examination of our present example we made a likely guess at a form of rule that would account for the finite protocol of states that we observed the system <math>X\!</math> passing through, as spied in the light of its boolean state variable <math>x : X \to \mathbb{B},\!</math> and that rule is well-formulated in any of these styles of notation:
 
 
On first examination of our present example we made a likely guess at a form of rule that
 
would account for the finite protocol of states that we observed the system <math>X\!</math> passing through, as spied in the light of its boolean state variable <math>x : X \to \mathbb{B},</math> and that rule is well-formulated in any of these styles of notation:
 
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
 
<math>\begin{array}{ll}
 
<math>\begin{array}{ll}
1.1. & f : \mathbb{B} \to \mathbb{B} ~\text{such that}~ f : \texttt{x} \mapsto \texttt{(x)}
+
1.1. & f : \mathbb{B} \to \mathbb{B} ~\text{such that}~ f : x \mapsto \texttt{(} x \texttt{)}
 
\\
 
\\
1.2. & \texttt{x}' ~=~ \texttt{(x)}
+
1.2. & x' ~=~ \texttt{(} x \texttt{)}
 
\\
 
\\
1.3. & \texttt{x} ~:=~ \texttt{(x)}
+
1.3. & x ~:=~ \texttt{(} x \texttt{)}
 
\\
 
\\
1.4. & \texttt{dx} ~=~ \texttt{1}
+
1.4. & \mathrm{d}x ~=~ 1
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
Line 623: Line 682:
 
|
 
|
 
<math>\begin{array}{ll}
 
<math>\begin{array}{ll}
2.1. & F : \mathbb{B}^2 \to \mathbb{B}^2 ~\text{such that}~ F : (\texttt{u}, \texttt{v}) \mapsto ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ )
+
2.1. & F : \mathbb{B}^2 \to \mathbb{B}^2 ~\text{such that}~ F : (u, v) \mapsto ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ )
 
\\
 
\\
2.2. & \texttt{u}' ~=~ \texttt{((u)(v))}~, \quad \texttt{v}' ~=~ \texttt{((u,~v))}
+
2.2. & u' ~=~ \texttt{((} u \texttt{)(} v \texttt{))} \quad ~,~ \quad v' ~=~ \texttt{((} u \texttt{,~} v \texttt{))}
 
\\
 
\\
2.3. & \texttt{u} ~:=~ \texttt{((u)(v))}~, \quad \texttt{v} ~:=~ \texttt{((u,~v))}
+
2.3. & u ~:=~ \texttt{((} u \texttt{)(} v \texttt{))} \quad ~,~ \quad v ~:=~ \texttt{((} u \texttt{,~} v \texttt{))}
 
\\
 
\\
2.4. & ???
+
2.4. & ?
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
Well, the last one is not such a fall off the log, but that is exactly the purpose for which we have been developing all of the foregoing machinations.
+
Well, the last one is not such a fall off a log, but that is exactly the purpose for which we have been developing all of the foregoing machinations.
  
 
Here is what I got when I just went ahead and calculated the finite differences willy-nilly:
 
Here is what I got when I just went ahead and calculated the finite differences willy-nilly:
Line 639: Line 698:
 
{| align="center" cellpadding="8" style="text-align:center"
 
{| align="center" cellpadding="8" style="text-align:center"
 
|-
 
|-
| <math>\text{Orbit 1. Intitial Point :}~ (u, v) = (1, 1)</math>
+
| <math>\text{Orbit 1. Intitial Point :}~ (u, v) = (1, 1)\!</math>
 
|-
 
|-
 
|
 
|
 
<math>\begin{array}{c|cc|cc|cc|cc|cc|c}
 
<math>\begin{array}{c|cc|cc|cc|cc|cc|c}
t & u & v & du & dv & d^2 u & d^2 v & d^3 u & d^3 v & d^4 u & d^4 v & \ldots \\
+
t & u & v & \mathrm{d}u & \mathrm{d}v & \mathrm{d}^2 u & \mathrm{d}^2 v & \mathrm{d}^3 u & \mathrm{d}^3 v & \mathrm{d}^4 u & \mathrm{d}^4 v & \ldots \\[8pt]
\\
+
0 & 1 & 1 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & \ldots \\
0 & 1 & 1 & 0 & 0 &     0 &     0 &     0 &     0 &     0 &     0 & \ldots \\
+
1 & 1 & 1 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & \ldots \\
1 & 1 & 1 & 0 & 0 &     0 &     0 &     0 &     0 &     0 &     0 & \ldots \\
+
2 & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & \ldots
4 & '' & '' & '' & '' &   '' &   '' &   '' &   '' &   '' &   '' & \ldots \\
 
 
\end{array}</math>
 
\end{array}</math>
 
|-
 
|-
| <math>\text{Orbit 2. Intitial Point :}~ (u, v) = (0, 0)</math>
+
| <math>\text{Orbit 2. Intitial Point :}~ (u, v) = (0, 0)\!</math>
 
|-
 
|-
 
|
 
|
 
<math>\begin{array}{c|cc|cc|cc|cc|cc|c}
 
<math>\begin{array}{c|cc|cc|cc|cc|cc|c}
t & u & v & du & dv & d^2 u & d^2 v & d^3 u & d^3 v & d^4 u & d^4 v & \ldots \\
+
t & u & v & \mathrm{d}u & \mathrm{d}v & \mathrm{d}^2 u & \mathrm{d}^2 v & \mathrm{d}^3 u & \mathrm{d}^3 v & \mathrm{d}^4 u & \mathrm{d}^4 v & \ldots \\[8pt]
\\
+
0 & 0 & 0 & 0 & 1 & 1 & 0 & 0 & 1 & 1 & 0 & \ldots \\
0 & 0 & 0 & 0 & 1 &     1 &     0 &     0 &     1 &     1 &     0 & \ldots \\
+
1 & 0 & 1 & 1 & 1 & 1 & 1 & 1 & 1 & 1 & 1 & \ldots \\
1 & 0 & 1 & 1 & 1 &     1 &     1 &     1 &     1 &     1 &     1 & \ldots \\
+
2 & 1 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & \ldots \\
2 & 1 & 0 & 0 & 0 &     0 &     0 &     0 &     0 &     0 &     0 & \ldots \\
+
3 & 1 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & \ldots \\
3 & 1 & 0 & 0 & 0 &     0 &     0 &     0 &     0 &     0 &     0 & \ldots \\
+
4 & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & \ldots
4 & '' & '' & '' & '' &   '' &   '' &   '' &   '' &   '' &   '' & \ldots \\
 
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
Line 668: Line 725:
 
What we are looking for is &mdash; one rule to rule them all, a rule that applies to every state and works every time.
 
What we are looking for is &mdash; one rule to rule them all, a rule that applies to every state and works every time.
  
What we see at first sight in the tables above are patterns of differential features that attach to the states in each orbit of the dynamics.  Looked at locally to these orbits, the isolated fixed point at <math>(1, 1)\!</math> is no problem, as the rule <math>\texttt{du~=~dv~=~0}</math> describes it pithily enough.  When it comes to the other orbit, the first thing that comes to mind is to write out the law <math>\texttt{du~=~v}, ~\texttt{dv~=~(u)}.</math>
+
What we see at first sight in the tables above are patterns of differential features that attach to the states in each orbit of the dynamics.  Looked at locally to these orbits, the isolated fixed point at <math>(1, 1)\!</math> is no problem, as the rule <math>\mathrm{d}u = \mathrm{d}v = 0\!</math> describes it pithily enough.  When it comes to the other orbit, the first thing that comes to mind is to write out the law <math>\mathrm{d}u = v ~,~ \mathrm{d}v = \texttt{(} u \texttt{)}.\!</math>
  
==Note 13==
+
==Symbolic Method==
  
 
It ought to be clear at this point that we need a more systematic symbolic method for computing the differentials of logical transformations, using the term ''differential'' in a loose way at present for all sorts of finite differences and derivatives, leaving it to another discussion to sharpen up its more exact technical senses.
 
It ought to be clear at this point that we need a more systematic symbolic method for computing the differentials of logical transformations, using the term ''differential'' in a loose way at present for all sorts of finite differences and derivatives, leaving it to another discussion to sharpen up its more exact technical senses.
Line 678: Line 735:
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{array}{lllll}
+
<math>\begin{matrix}
F & = & (f, g) & = & ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ ).
+
F & = & (f, g) & = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ ).
\end{array}</math>
+
\end{matrix}</math>
 
|}
 
|}
  
In their application to this logical transformation the operators <math>\operatorname{E}</math> and <math>\operatorname{D}</math> respectively produce the ''enlarged map'' <math>\operatorname{E}F = (\operatorname{E}f, \operatorname{E}g)</math> and the ''difference map'' <math>\operatorname{D}F = (\operatorname{D}f, \operatorname{D}g),</math> whose components can be given as follows, if the reader, in the absence of a special format for logical parentheses, can forgive syntactically bilingual phrases:
+
In their application to this logical transformation the operators <math>\mathrm{E}\!</math> and <math>\mathrm{D}\!</math> respectively produce the ''enlarged map'' <math>\mathrm{E}F = (\mathrm{E}f, \mathrm{E}g)\!</math> and the ''difference map'' <math>\mathrm{D}F = (\mathrm{D}f, \mathrm{D}g),\!</math> whose components can be given as follows.
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
 
<math>\begin{array}{lll}
 
<math>\begin{array}{lll}
\operatorname{E}f & = & \texttt{(( u + du )( v + dv ))}
+
\mathrm{E}f & = & \texttt{((} u + \mathrm{d}u \texttt{)(} v + \mathrm{d}v \texttt{))}
\\ \\
+
\\[8pt]
\operatorname{E}g & = & \texttt{(( u + du ,~ v + dv ))}
+
\mathrm{E}g & = & \texttt{((} u + \mathrm{d}u \texttt{,~} v + \mathrm{d}v \texttt{))}
\\ \\
+
\\[8pt]
\operatorname{D}f & = & \texttt{((u)(v)) ~+~ (( u + du )( v + dv ))}
+
\mathrm{D}f & = & \texttt{((} u \texttt{)(} v \texttt{))} ~+~ \texttt{((} u + \mathrm{d}u \texttt{)(} v + \mathrm{d}v \texttt{))}
\\ \\
+
\\[8pt]
\operatorname{D}g & = & \texttt{((u,~v)) ~+~ (( u + du ,~ v + dv ))}
+
\mathrm{D}g & = & \texttt{((} u \texttt{,~} v \texttt{))} ~+~ \texttt{((} u + \mathrm{d}u \texttt{,~} v + \mathrm{d}v \texttt{))}
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
Line 700: Line 757:
 
But these initial formulas are purely definitional, and help us little to understand either the purpose of the operators or the significance of the results.  Working symbolically, let's apply a more systematic method to the separate components of the mapping <math>F.\!</math>
 
But these initial formulas are purely definitional, and help us little to understand either the purpose of the operators or the significance of the results.  Working symbolically, let's apply a more systematic method to the separate components of the mapping <math>F.\!</math>
  
A sketch of this work is presented in the following series of Figures, where each logical proposition is expanded over the basic cells <math>\texttt{uv}, \texttt{u(v)}, \texttt{(u)v}, \texttt{(u)(v)}</math> of the 2-dimensional universe of discourse <math>U^\circ = [u, v].\!</math>
+
A sketch of this work is presented in the following series of Figures, where each logical proposition is expanded over the basic cells <math>uv, u \texttt{(} v \texttt{)}, \texttt{(} u \texttt{)} v, \texttt{(} u \texttt{)(} v \texttt{)}\!</math> of the 2-dimensional universe of discourse <math>U^\bullet = [u, v].\!</math>
  
===Computation Summary : <math>f(u, v) = \texttt{((u)(v))}</math>===
+
===Computation Summary for Logical Disjunction===
  
The venn diagram in Figure&nbsp;1.1 shows how the proposition <math>f = \texttt{((u)(v))}</math> can be expanded over the universe of discourse <math>[u, v]\!</math> to produce a logically equivalent exclusive disjunction, namely, <math>\texttt{uv~+~u(v)~+~(u)v}.</math>
+
The venn diagram in Figure&nbsp;1.1 shows how the proposition <math>f = \texttt{((} u \texttt{)(} v \texttt{))}\!</math> can be expanded over the universe of discourse <math>[u, v]\!</math> to produce a logically equivalent exclusive disjunction, namely, <math>uv + u \texttt{(} v \texttt{)} + \texttt{(} u \texttt{)} v.\!</math>
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 746: Line 805:
 
Figure 1.1.  f = ((u)(v))
 
Figure 1.1.  f = ((u)(v))
 
</pre>
 
</pre>
 +
|}
  
Figure&nbsp;1.2 expands <math>\operatorname{E}f = \texttt{((u + du)(v + dv))}</math> over <math>[u, v]\!</math> to give:
+
Figure&nbsp;1.2 expands <math>\mathrm{E}f = \texttt{((} u + \mathrm{d}u \texttt{)(} v + \mathrm{d}v \texttt{))}\!</math> over <math>[u, v]\!</math> to give:
  
{| align="center" cellpadding="8" width="90%"
+
{| align="center" cellpadding="8" style="text-align:center; width:100%"
| <math>\texttt{uv~(du~dv) ~+~ u(v)~(du (dv)) ~+~ (u)v~((du) dv) ~+~ (u)(v)~((du)(dv))}</math>
+
|
 +
<math>\begin{matrix}
 +
\mathrm{E}\texttt{((} u \texttt{)(} v \texttt{))}
 +
& = & uv \cdot \texttt{(} \mathrm{d}u ~ \mathrm{d}v \texttt{)}
 +
& + & u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{(} \mathrm{d}v \texttt{))}
 +
& + & \texttt{(} u \texttt{)} v \cdot \texttt{((} \mathrm{d}u \texttt{)} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{((} \mathrm{d}u \texttt{)(} \mathrm{d}v \texttt{))}
 +
\end{matrix}</math>
 
|}
 
|}
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 793: Line 862:
 
Figure 1.2.  Ef = ((u + du)(v + dv))
 
Figure 1.2.  Ef = ((u + du)(v + dv))
 
</pre>
 
</pre>
 +
|}
  
Figure&nbsp;1.3 expands <math>\operatorname{D}f = f + \operatorname{E}f</math> over <math>[u, v]\!</math> to produce:
+
Figure&nbsp;1.3 expands <math>\mathrm{D}f = f + \mathrm{E}f\!</math> over <math>[u, v]\!</math> to produce:
  
{| align="center" cellpadding="8" width="90%"
+
{| align="center" cellpadding="8" style="text-align:center; width:100%"
| <math>\texttt{uv~du~dv ~+~ u(v)~du(dv) ~+~ (u)v~(du)dv ~+~ (u)(v)~((du)(dv))}</math>
+
|
|}
+
<math>\begin{matrix}
 +
\mathrm{D}\texttt{((} u \texttt{)(} v \texttt{))}
 +
& = & uv \cdot \mathrm{d}u ~ \mathrm{d}v
 +
& + & u \texttt{(} v \texttt{)} \cdot \mathrm{d}u \texttt{(} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{)} \mathrm{d}v
 +
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{((} \mathrm{d}u \texttt{)(} \mathrm{d}v \texttt{))}
 +
\end{matrix}</math>
 +
|}
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 840: Line 919:
 
Figure 1.3.  Df = f + Ef
 
Figure 1.3.  Df = f + Ef
 
</pre>
 
</pre>
 +
|}
  
 
I'll break this here in case anyone wants to try and do the work for <math>g\!</math> on their own.
 
I'll break this here in case anyone wants to try and do the work for <math>g\!</math> on their own.
  
==Note 14==
+
===Computation Summary for Logical Equality===
  
===Computation Summary : <math>g(u, v) = \texttt{((u,~v))}</math>===
+
The venn diagram in Figure&nbsp;2.1 shows how the proposition <math>g = \texttt{((} u \texttt{,~} v \texttt{))}\!</math> can be expanded over the universe of discourse <math>[u, v]\!</math> to produce a logically equivalent exclusive disjunction, namely, <math>uv + \texttt{(} u \texttt{)(} v \texttt{)}.\!</math>
 
 
The venn diagram in Figure&nbsp;2.1 shows how the proposition <math>g = \texttt{((u,~v))}</math> can be expanded over the universe of discourse <math>[u, v]\!</math> to produce a logically equivalent exclusive disjunction, namely, <math>\texttt{uv ~+~ (u)(v)}.</math>
 
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 889: Line 969:
 
Figure 2.1.  g = ((u, v))
 
Figure 2.1.  g = ((u, v))
 
</pre>
 
</pre>
 +
|}
  
Figure&nbsp;2.2 expands <math>\operatorname{E}g = \texttt{((u + du,~v + dv))}</math> over <math>[u, v]\!</math> to give:
+
Figure&nbsp;2.2 expands <math>\mathrm{E}g = \texttt{((} u + \mathrm{d}u \texttt{,~} v + \mathrm{d}v \texttt{))}\!</math> over <math>[u, v]\!</math> to give:
  
{| align="center" cellpadding="8" width="90%"
+
{| align="center" cellpadding="8" style="text-align:center; width:100%"
| <math>\texttt{uv~((du,~dv)) ~+~ u(v)~(du,~dv) ~+~ (u)v~(du,~dv) ~+~ (u)(v)~((du,~dv))}</math>
+
|
 +
<math>\begin{matrix}
 +
\mathrm{E}\texttt{((} u \texttt{,~} v \texttt{))}
 +
& = & uv \cdot \texttt{((} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{))}
 +
& + & u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{((} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{))}
 +
\end{matrix}</math>
 
|}
 
|}
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 936: Line 1,026:
 
Figure 2.2.  Eg = ((u + du, v + dv))
 
Figure 2.2.  Eg = ((u + du, v + dv))
 
</pre>
 
</pre>
 +
|}
  
Figure&nbsp;2.3 expands <math>\operatorname{D}g = g + \operatorname{E}g</math> over <math>[u, v]\!</math> to yield the form:
+
Figure&nbsp;2.3 expands <math>\mathrm{D}g = g + \mathrm{E}g\!</math> over <math>[u, v]\!</math> to yield the form:
  
{| align="center" cellpadding="8" width="90%"
+
{| align="center" cellpadding="8" style="text-align:center; width:100%"
| <math>\texttt{uv~(du,~dv) ~+~ u(v)~(du,~dv) ~+~ (u)v~(du,~dv) ~+~ (u)(v)~(du,~dv)}</math>
+
|
 +
<math>\begin{matrix}
 +
\mathrm{D}\texttt{((} u \texttt{,~} v \texttt{))}
 +
& = & uv \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
& + & u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
\end{matrix}</math>
 
|}
 
|}
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 983: Line 1,083:
 
Figure 2.3.  Dg = g + Eg
 
Figure 2.3.  Dg = g + Eg
 
</pre>
 
</pre>
 +
|}
  
==Note 15==
+
==Differential : Locally Linear Approximation==
  
 
{| cellpadding="2" cellspacing="2" width="100%"
 
{| cellpadding="2" cellspacing="2" width="100%"
Line 1,005: Line 1,106:
 
|
 
|
 
<math>\begin{array}{lllll}
 
<math>\begin{array}{lllll}
F & = & (f, g) & = & ( ~\texttt{((u)(v))}~ , ~\texttt{((u,~v))}~ ).
+
F
 +
& = & (f, g)
 +
& = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ )
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
To speed things along, I will skip a mass of motivating discussion and just exhibit the simplest form of a differential <math>\operatorname{d}F\!</math> for the current example of a logical transformation <math>F,\!</math> after which the majority of the easiest questions will have been answered in visually intuitive terms.
+
To speed things along, I will skip a mass of motivating discussion and just exhibit the simplest form of a differential <math>\mathrm{d}F\!</math> for the current example of a logical transformation <math>F,\!</math> after which the majority of the easiest questions will have been answered in visually intuitive terms.
  
For <math>F = (f, g)\!</math> we have <math>\operatorname{d}F = (\operatorname{d}f, \operatorname{d}g),</math> and so we can proceed componentwise, patching the pieces back together at the end.
+
For <math>F = (f, g)\!</math> we have <math>\mathrm{d}F = (\mathrm{d}f, \mathrm{d}g),\!</math> and so we can proceed componentwise, patching the pieces back together at the end.
  
 
We have prepared the ground already by computing these terms:
 
We have prepared the ground already by computing these terms:
Line 1,018: Line 1,121:
 
|
 
|
 
<math>\begin{array}{lll}
 
<math>\begin{array}{lll}
\operatorname{E}f & = & \texttt{(( u + du )( v + dv ))}
+
\mathrm{E}f & = & \texttt{((} u + \mathrm{d}u \texttt{)(} v + \mathrm{d}v \texttt{))}
\\ \\
+
\\[8pt]
\operatorname{E}g & = & \texttt{(( u + du ,~ v + dv ))}
+
\mathrm{E}g & = & \texttt{((} u + \mathrm{d}u \texttt{,~} v + \mathrm{d}v \texttt{))}
\\ \\
+
\\[8pt]
\operatorname{D}f & = & \texttt{((u)(v)) ~+~ (( u + du )( v + dv ))}
+
\mathrm{D}f & = & \texttt{((} u \texttt{)(} v \texttt{))} ~+~ \texttt{((} u + \mathrm{d}u \texttt{)(} v + \mathrm{d}v \texttt{))}
\\ \\
+
\\[8pt]
\operatorname{D}g & = & \texttt{((u,~v)) ~+~ (( u + du ,~ v + dv ))}
+
\mathrm{D}g & = & \texttt{((} u \texttt{,~} v \texttt{))} ~+~ \texttt{((} u + \mathrm{d}u \texttt{,~} v + \mathrm{d}v \texttt{))}
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
As a matter of fact, computing the symmetric differences <math>\operatorname{D}f = f + \operatorname{E}f</math> and <math>\operatorname{D}g = g + \operatorname{E}g</math> has already taken care of the ''localizing'' part of the task by subtracting out the forms of <math>f\!</math> and <math>g\!</math> from the forms of <math>\operatorname{E}f</math> and <math>\operatorname{E}g,</math> respectively.  Thus all we have left to do is to decide what linear propositions best approximate the difference maps <math>\operatorname{D}f</math> and <math>\operatorname{D}g,</math> respectively.
+
As a matter of fact, computing the symmetric differences <math>\mathrm{D}f = f + \mathrm{E}f\!</math> and <math>\mathrm{D}g = g + \mathrm{E}g\!</math> has already taken care of the ''localizing'' part of the task by subtracting out the forms of <math>f\!</math> and <math>g\!</math> from the forms of <math>\mathrm{E}f\!</math> and <math>\mathrm{E}g,\!</math> respectively.  Thus all we have left to do is to decide what linear propositions best approximate the difference maps <math>{\mathrm{D}f}\!</math> and <math>{\mathrm{D}g},\!</math> respectively.
  
 
This raises the question:  What is a linear proposition?
 
This raises the question:  What is a linear proposition?
  
The answer that makes the most sense in this context is this:  A proposition is just a boolean-valued function, so a linear proposition is a linear function into the boolean space <math>\mathbb{B}.</math>
+
The answer that makes the most sense in this context is this:  A proposition is just a boolean-valued function, so a linear proposition is a linear function into the boolean space <math>\mathbb{B}.\!</math>
  
In particular, the linear functions that we want will be linear functions in the differential variables <math>du\!</math> and <math>dv.\!</math>
+
In particular, the linear functions that we want will be linear functions in the differential variables <math>\mathrm{d}u\!</math> and <math>\mathrm{d}v.\!</math>
  
As it turns out, there are just four linear propositions in the associated ''differential universe'' <math>\operatorname{d}U^\circ = [du, dv],</math> and these are the propositions that are commonly denoted:  <math>\texttt{0}, \texttt{du}, \texttt{dv}, \texttt{du + dv},</math> in other words, <math>\texttt{()}, \texttt{du}, \texttt{dv}, \texttt{(du, dv)}.</math>
+
As it turns out, there are just four linear propositions in the associated ''differential universe'' <math>\mathrm{d}U^\bullet = [\mathrm{d}u, \mathrm{d}v].\!</math> These are the propositions that are commonly denoted:  <math>{0, ~\mathrm{d}u, ~\mathrm{d}v, ~\mathrm{d}u + \mathrm{d}v},\!</math> in other words<math>\texttt{(~)}, ~\mathrm{d}u, ~\mathrm{d}v, ~\texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}.\!</math>
  
==Note 16==
+
==Notions of Approximation==
  
 
{| cellpadding="2" cellspacing="2" width="100%"
 
{| cellpadding="2" cellspacing="2" width="100%"
Line 1,063: Line 1,166:
 
Justifying a notion of approximation is a little more involved in general, and especially in these discrete logical spaces, than it would be expedient for people in a hurry to tangle with right now.  I will just say that there are ''naive'' or ''obvious'' notions and there are ''sophisticated'' or ''subtle'' notions that we might choose among.  The later would engage us in trying to construct proper logical analogues of Lie derivatives, and so let's save that for when we have become subtle or sophisticated or both.  Against or toward that day, as you wish, let's begin with an option in plain view.
 
Justifying a notion of approximation is a little more involved in general, and especially in these discrete logical spaces, than it would be expedient for people in a hurry to tangle with right now.  I will just say that there are ''naive'' or ''obvious'' notions and there are ''sophisticated'' or ''subtle'' notions that we might choose among.  The later would engage us in trying to construct proper logical analogues of Lie derivatives, and so let's save that for when we have become subtle or sophisticated or both.  Against or toward that day, as you wish, let's begin with an option in plain view.
  
Figure&nbsp;1.4 illustrates one way of ranging over the cells of the underlying universe <math>U^\circ = [u, v]\!</math> and selecting at each cell the linear proposition in <math>\operatorname{d}U^\circ = [du, dv]</math> that best approximates the patch of the difference map <math>\operatorname{D}f</math> that is located there, yielding the following formula for the differential <math>\operatorname{d}f.</math>
+
Figure&nbsp;1.4 illustrates one way of ranging over the cells of the underlying universe <math>U^\bullet = [u, v]\!</math> and selecting at each cell the linear proposition in <math>\mathrm{d}U^\bullet = [\mathrm{d}u, \mathrm{d}v]\!</math> that best approximates the patch of the difference map <math>{\mathrm{D}f}\!</math> that is located there, yielding the following formula for the differential <math>\mathrm{d}f.\!</math>
  
{| align="center" cellpadding="8" width="90%"
+
{| align="center" cellpadding="8" style="text-align:center; width:100%"
| <math>\operatorname{d}f ~=~ \texttt{uv} \cdot \texttt{0} ~+~ \texttt{u(v)} \cdot \texttt{du} ~+~ \texttt{(u)v} \cdot \texttt{dv} ~+~ \texttt{(u)(v)} \cdot \texttt{(du, dv)}</math>
+
|
 +
<math>\begin{array}{*{11}{c}}
 +
\mathrm{d}f
 +
& = & \mathrm{d}\texttt{((} u \texttt{)(} v \texttt{))}
 +
& = & uv \cdot 0
 +
& + & u \texttt{(} v \texttt{)} \cdot \mathrm{d}u
 +
& + & \texttt{(} u \texttt{)} v \cdot \mathrm{d}v
 +
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
\end{array}</math>
 
|}
 
|}
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,109: Line 1,222:
 
Figure 1.4.  df = linear approx to Df
 
Figure 1.4.  df = linear approx to Df
 
</pre>
 
</pre>
 +
|}
  
Figure&nbsp;2.4 illustrates one way of ranging over the cells of the underlying universe <math>U^\circ = [u, v]\!</math> and selecting at each cell the linear proposition in <math>\operatorname{d}U^\circ = [du, dv]</math> that best approximates the patch of the difference map <math>\operatorname{D}g</math> that is located there, yielding the following formula for the differential <math>\operatorname{d}g.</math>
+
Figure&nbsp;2.4 illustrates one way of ranging over the cells of the underlying universe <math>U^\bullet = [u, v]\!</math> and selecting at each cell the linear proposition in <math>\mathrm{d}U^\bullet = [du, dv]\!</math> that best approximates the patch of the difference map <math>\mathrm{D}g\!</math> that is located there, yielding the following formula for the differential <math>\mathrm{d}g.\!</math>
  
{| align="center" cellpadding="8" width="90%"
+
{| align="center" cellpadding="8" style="text-align:center; width:100%"
| <math>\operatorname{d}g ~=~ \texttt{uv} \cdot \texttt{(du, dv)} ~+~ \texttt{u(v)} \cdot \texttt{(du, dv)} ~+~ \texttt{(u)v} \cdot \texttt{(du, dv)} ~+~ \texttt{(u)(v)} \cdot \texttt{(du, dv)}</math>
+
|
 +
<math>\begin{array}{*{11}{c}}
 +
\mathrm{d}g
 +
& = & \mathrm{d}\texttt{((} u \texttt{,} v \texttt{))}
 +
& = & uv \cdot \texttt{(} \mathrm{d}u \texttt{,} \mathrm{d}v \texttt{)}
 +
& + & u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{,} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,} \mathrm{d}v \texttt{)}
 +
\end{array}</math>
 
|}
 
|}
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,156: Line 1,280:
 
Figure 2.4.  dg = linear approx to Dg
 
Figure 2.4.  dg = linear approx to Dg
 
</pre>
 
</pre>
 +
|}
  
Well, <math>g,\!</math> that was easy, seeing as how <math>\operatorname{D}g</math> is already linear at each locus, <math>\operatorname{d}g = \operatorname{D}g.</math>
+
Well, <math>g,\!</math> that was easy, seeing as how <math>\mathrm{D}g\!</math> is already linear at each locus, <math>\mathrm{d}g = \mathrm{D}g.\!</math>
  
==Note 17==
+
==Analytic Series==
  
We have been conducting the differential analysis of the logical transformation <math>F : [u, v] \mapsto [u, v]</math> defined as <math>F : (u, v) \mapsto ( ~\texttt{((u)(v))}~, ~\texttt{((u, v))}~ ),</math> and this means starting with the extended transformation <math>\operatorname{E}F : [u, v, du, dv] \to [u, v, du, dv]</math> and breaking it into an analytic series, <math>\operatorname{E}F = F + \operatorname{d}F + \operatorname{d}^2 F + \ldots,</math> and
+
We have been conducting the differential analysis of the logical transformation <math>F : [u, v] \mapsto [u, v]\!</math> defined as <math>F : (u, v) \mapsto ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ ),\!</math> and this means starting with the extended transformation <math>\mathrm{E}F : [u, v, \mathrm{d}u, \mathrm{d}v] \to [u, v, \mathrm{d}u, \mathrm{d}v]\!</math> and breaking it into an analytic series, <math>\mathrm{E}F = F + \mathrm{d}F + \mathrm{d}^2 F + \ldots,\!</math> and so on until there is nothing left to analyze any further.
so on until there is nothing left to analyze any further.
 
  
 
As a general rule, one proceeds by way of the following stages:
 
As a general rule, one proceeds by way of the following stages:
Line 1,168: Line 1,292:
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{array}{lccccc}
+
<math>\begin{array}{*{6}{l}}
1. & \operatorname{E}F   & = & \operatorname{d}^0 F & + & \operatorname{r}^0 F
+
1. & \mathrm{E}F
 +
& = & \mathrm{d}^0 F
 +
& + & \mathrm{r}^0 F
 
\\
 
\\
2. & \operatorname{r}^0 F & = & \operatorname{d}^1 F & + & \operatorname{r}^1 F
+
2. & \mathrm{r}^0 F
 +
& = & \mathrm{d}^1 F
 +
& + & \mathrm{r}^1 F
 
\\
 
\\
3. & \operatorname{r}^1 F & = & \operatorname{d}^2 F & + & \operatorname{r}^2 F
+
3. & \mathrm{r}^1 F
 +
& = & \mathrm{d}^2 F
 +
& + & \mathrm{r}^2 F
 
\\
 
\\
 
4. & \ldots
 
4. & \ldots
Line 1,179: Line 1,309:
 
|}
 
|}
  
In our analysis of the transformation <math>F,\!</math> we carried out Step&nbsp;1 in the more familiar form <math>\operatorname{E}F = F + \operatorname{D}F,</math> and we have just reached Step&nbsp;2 in the form <math>\operatorname{D}F = \operatorname{d}F + \operatorname{r}F,</math> where <math>\operatorname{r}F</math> is the residual term that remains for us to examine next.
+
In our analysis of the transformation <math>F,\!</math> we carried out Step&nbsp;1 in the more familiar form <math>\mathrm{E}F = F + \mathrm{D}F\!</math> and we have just reached Step&nbsp;2 in the form <math>\mathrm{D}F = \mathrm{d}F + \mathrm{r}F,\!</math> where <math>\mathrm{r}F\!</math> is the residual term that remains for us to examine next.
 
 
'''NB.'''  I'm am trying to give quick overview here, and this forces me to omit many picky details.  The picky reader may wish to consult the more detailed presentation of this material at the following locations:
 
  
; Jon Awbrey, "[[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0|Differential Logic and Dynamic Systems]]"
+
'''Note.'''  I'm am trying to give quick overview here, and this forces me to omit many picky details. The picky reader may wish to consult the more detailed presentation of this material at the following locations:
: [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0#The_Secant_Operator_:_E|The Secant Operator]]
 
: [[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0#Taking_Aim_at_Higher_Dimensional_Targets|Higher Dimensional Targets]]
 
  
==Note 18==
+
:* [[Differential Logic and Dynamic Systems 2.0|Differential Logic and Dynamic Systems]]
 +
::* [[Differential Logic and Dynamic Systems 2.0#The Secant Operator : E|The Secant Operator]]
 +
::* [[Differential Logic and Dynamic Systems 2.0#Taking Aim at Higher Dimensional Targets|Higher Dimensional Targets]]
  
 
Let's push on with the analysis of the transformation:
 
Let's push on with the analysis of the transformation:
Line 1,194: Line 1,322:
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
F & : & (u, v) & \mapsto & (f(u, v),~g(u, v)) & = & (~\texttt{((u)(v))}~,~\texttt{((u,~v))}~).\end{matrix}</math>
+
F & : & (u, v) & \mapsto & (f(u, v), g(u, v))
 +
& = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~)
 +
\end{matrix}</math>
 
|}
 
|}
  
 
For ease of comparison and computation, I will collect the Figures that we need for the remainder of the work together on one page.
 
For ease of comparison and computation, I will collect the Figures that we need for the remainder of the work together on one page.
  
===Computation Summary : <math>f(u, v) = \texttt{((u)(v))}</math>===
+
===Computation Summary for Logical Disjunction===
  
Figure&nbsp;1.1 shows the expansion of <math>f = \texttt{((u)(v))}</math> over <math>[u, v]\!</math> to produce the expression:
+
Figure&nbsp;1.1 shows the expansion of <math>f = \texttt{((} u \texttt{)(} v \texttt{))}\!</math> over <math>[u, v]\!</math> to produce the expression:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
| <math>\texttt{uv} ~+~ \texttt{u(v)} ~+~ \texttt{(u)v}</math>
+
|
 +
<math>\begin{matrix}
 +
uv & + & u \texttt{(} v \texttt{)} & + & \texttt{(} u \texttt{)} v
 +
\end{matrix}</math>
 
|}
 
|}
  
Figure&nbsp;1.2 shows the expansion of <math>\operatorname{E}f = \texttt{((u + du)(v + dv))}</math> over <math>[u, v]\!</math> to produce the expression:
+
Figure&nbsp;1.2 shows the expansion of <math>\mathrm{E}f = \texttt{((} u + \mathrm{d}u \texttt{)(} v + \mathrm{d}v \texttt{))}\!</math> over <math>[u, v]\!</math> to produce the expression:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
| <math>\texttt{uv} \cdot \texttt{(du~dv)} + \texttt{u(v)} \cdot \texttt{(du (dv))} + \texttt{(u)v} \cdot \texttt{((du) dv)} + \texttt{(u)(v)} \cdot \texttt{((du)(dv))}</math>
+
|
 +
<math>\begin{matrix}
 +
uv \cdot \texttt{(} \mathrm{d}u ~ \mathrm{d}v \texttt{)} & + &
 +
u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{(} \mathrm{d}v \texttt{))} & + &
 +
\texttt{(} u \texttt{)} v \cdot \texttt{((} \mathrm{d}u \texttt{)} \mathrm{d}v \texttt{)} & + &
 +
\texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{((} \mathrm{d}u \texttt{)(} \mathrm{d}v \texttt{))}
 +
\end{matrix}</math>
 
|}
 
|}
  
<math>\operatorname{E}f</math> tells you what you would have to do, from where you are in the universe <math>[u, v],\!</math> if you want to end up in a place where <math>f\!</math> is true.  In this case, where the prevailing proposition <math>f\!</math> is <math>\texttt{((u)(v))},</math> the indication <math>\texttt{uv} \cdot \texttt{(du~dv)}</math> of <math>\operatorname{E}f</math> tells you this:  If <math>u\!</math> and <math>v\!</math> are both true where you are, then just don't change both <math>u\!</math> and <math>v\!</math> at once, and you will end up in a place where <math>\texttt{((u)(v))}</math> is true.
+
In general, <math>\mathrm{E}f\!</math> tells you what you would have to do, from wherever you are in the universe <math>[u, v],\!</math> if you want to end up in a place where <math>f\!</math> is true.  In this case, where the prevailing proposition <math>f\!</math> is <math>\texttt{((} u \texttt{)(} v \texttt{))},\!</math> the indication <math>uv \cdot \texttt{(} \mathrm{d}u ~ \mathrm{d}v \texttt{)}\!</math> of <math>\mathrm{E}f\!</math> tells you this:  If <math>u\!</math> and <math>v\!</math> are both true where you are, then just don't change both <math>u\!</math> and <math>v\!</math> at once, and you will end up in a place where <math>\texttt{((} u \texttt{)(} v \texttt{))}\!</math> is true.
  
Figure&nbsp;1.3 shows the expansion of <math>\operatorname{D}f</math> over <math>[u, v]\!</math> to produce the expression:
+
Figure&nbsp;1.3 shows the expansion of <math>\mathrm{D}f</math> over <math>[u, v]\!</math> to produce the expression:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
| <math>\texttt{uv} \cdot \texttt{du~dv} ~+~ \texttt{u(v)} \cdot \texttt{du(dv)} ~+~ \texttt{(u)v} \cdot \texttt{(du)dv} ~+~ \texttt{(u)(v)} \cdot \texttt{((du)(dv))}</math>
+
|
 +
<math>\begin{matrix}
 +
uv \cdot \mathrm{d}u ~ \mathrm{d}v & + &
 +
u \texttt{(} v \texttt{)} \cdot \mathrm{d}u \texttt{(} \mathrm{d}v \texttt{)} & + &
 +
\texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{)} \mathrm{d}v & + &
 +
\texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{((} \mathrm{d}u \texttt{)(} \mathrm{d}v \texttt{))}
 +
\end{matrix}</math>
 
|}
 
|}
  
<math>\operatorname{D}f</math> tells you what you would have to do, from where you are in the universe <math>[u, v],\!</math> if you want to bring about a change in the value of <math>f,\!</math> that is, if you want to get to a place where the value of <math>f\!</math> is different from what it is where you are.  In the present case, where the reigning proposition <math>f\!</math> is <math>\texttt{((u)(v))},</math> the term <math>\texttt{uv} \cdot \texttt{du~dv}</math> of <math>\operatorname{D}f</math> tells you this:  If <math>u\!</math> and <math>v\!</math> are both true where you are, then you would have to change both <math>u\!</math> and <math>v\!</math> in order to reach a place where the value of <math>f\!</math> is different from what it is where you are.
+
In general, <math>{\mathrm{D}f}\!</math> tells you what you would have to do, from wherever you are in the universe <math>[u, v],\!</math> if you want to bring about a change in the value of <math>f,\!</math> that is, if you want to get to a place where the value of <math>f\!</math> is different from what it is where you are.  In the present case, where the reigning proposition <math>f\!</math> is <math>\texttt{((} u \texttt{)(} v \texttt{))},\!</math> the term <math>uv \cdot \mathrm{d}u ~ \mathrm{d}v\!</math> of <math>{\mathrm{D}f}\!</math> tells you this:  If <math>u\!</math> and <math>v\!</math> are both true where you are, then you would have to change both <math>u\!</math> and <math>v\!</math> in order to reach a place where the value of <math>f\!</math> is different from what it is where you are.
  
Figure&nbsp;1.4 approximates <math>\operatorname{D}f</math> by the linear form <math>\operatorname{d}f</math> that expands over <math>[u, v]\!</math> as follows:
+
Figure&nbsp;1.4 approximates <math>{\mathrm{D}f}\!</math> by the linear form <math>\mathrm{d}f\!</math> that expands over <math>[u, v]\!</math> as follows:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
 
<math>\begin{matrix}
 
<math>\begin{matrix}
\operatorname{d}f
+
\mathrm{d}f
& = & \texttt{uv} \cdot \texttt{0} & + & \texttt{u(v)} \cdot \texttt{du} & + & \texttt{(u)v} \cdot \texttt{dv} & + & \texttt{(u)(v)} \cdot \texttt{(du, dv)}
+
& = & uv \cdot 0
\\ \\
+
& + & u \texttt{(} v \texttt{)} \cdot \mathrm{d}u
& = & & & \texttt{u(v)} \cdot \texttt{du} & + & \texttt{(u)v} \cdot \texttt{dv} & + & \texttt{(u)(v)} \cdot \texttt{(du, dv)}
+
& + & \texttt{(} u \texttt{)} v \cdot \mathrm{d}v
 +
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
\\[8pt]
 +
& = &&& u \texttt{(} v \texttt{)} \cdot \mathrm{d}u
 +
& + &   \texttt{(} u \texttt{)} v \cdot \mathrm{d}v
 +
& + &   \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|}
 
|}
  
Figure&nbsp;1.5 shows what remains of the difference map <math>\operatorname{D}f</math> when the first order linear contribution <math>\operatorname{d}f</math> is removed, namely:
+
Figure&nbsp;1.5 shows what remains of the difference map <math>{\mathrm{D}f}\!</math> when the first order linear contribution <math>\mathrm{d}f\!</math> is removed, namely:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{matrix}
+
<math>\begin{array}{*{9}{l}}
\operatorname{r}f
+
\mathrm{r}f
& = & \texttt{uv} \cdot \texttt{du~dv} & + & \texttt{u(v)} \cdot \texttt{du~dv} & + & \texttt{(u)v} \cdot \texttt{du~dv} & + & \texttt{(u)(v)} \cdot \texttt{du~dv}
+
& = & uv \cdot \mathrm{d}u ~ \mathrm{d}v
\\ \\
+
& + & u \texttt{(} v \texttt{)} \cdot \mathrm{d}u ~ \mathrm{d}v
& = & \texttt{du~dv}
+
& + & \texttt{(} u \texttt{)} v \cdot \mathrm{d}u ~ \mathrm{d}v
\end{matrix}</math>
+
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \mathrm{d}u ~ \mathrm{d}v
 +
\\[8pt]
 +
& = & \mathrm{d}u ~ \mathrm{d}v
 +
\end{array}</math>
 
|}
 
|}
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,287: Line 1,442:
 
Figure 1.1.  f = ((u)(v))
 
Figure 1.1.  f = ((u)(v))
 
</pre>
 
</pre>
 +
|}
  
<br>
+
{| align="center" border="0" cellpadding="10"
 
+
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,330: Line 1,486:
 
Figure 1.2.  Ef = ((u + du)(v + dv))
 
Figure 1.2.  Ef = ((u + du)(v + dv))
 
</pre>
 
</pre>
 +
|}
  
<br>
+
{| align="center" border="0" cellpadding="10"
 
+
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,373: Line 1,530:
 
Figure 1.3.  Difference Map Df = f + Ef
 
Figure 1.3.  Difference Map Df = f + Ef
 
</pre>
 
</pre>
 +
|}
  
<br>
+
{| align="center" border="0" cellpadding="10"
 
+
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,416: Line 1,574:
 
Figure 1.4.  Linear Proxy df for Df
 
Figure 1.4.  Linear Proxy df for Df
 
</pre>
 
</pre>
 +
|}
  
<br>
+
{| align="center" border="0" cellpadding="10"
 
+
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,459: Line 1,618:
 
Figure 1.5.  Remainder rf = Df + df
 
Figure 1.5.  Remainder rf = Df + df
 
</pre>
 
</pre>
 +
|}
  
==Note 19==
+
===Computation Summary for Logical Equality===
 
 
===Computation Summary : <math>g(u, v) = \texttt{((u, v))}</math>===
 
  
Figure&nbsp;2.1 shows the expansion of <math>g = \texttt{((u, v))}</math> over <math>[u, v]\!</math> to produce the expression:
+
Figure&nbsp;2.1 shows the expansion of <math>g = \texttt{((} u \texttt{,~} v \texttt{))}\!</math> over <math>[u, v]\!</math> to produce the expression:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
| <math>\texttt{uv} ~+~ \texttt{(u)(v)}</math>
+
|
 +
<math>\begin{matrix}
 +
uv & + & \texttt{(} u \texttt{)(} v \texttt{)}
 +
\end{matrix}</math>
 
|}
 
|}
  
Figure&nbsp;2.2 shows the expansion of <math>\operatorname{E}g = \texttt{((u + du, v + dv))}</math> over <math>[u, v]\!</math> to produce the expression:
+
Figure&nbsp;2.2 shows the expansion of <math>\mathrm{E}g = \texttt{((} u + \mathrm{d}u \texttt{,~} v + \mathrm{d}v \texttt{))}\!</math> over <math>[u, v]\!</math> to produce the expression:
  
{| align="center" cellpadding="8" width="90%"
+
{| align="center" cellpadding="8" width="90%"
| <math>\texttt{uv} \cdot \texttt{((du, dv))} + \texttt{u(v)} \cdot \texttt{(du, dv)} + \texttt{(u)v} \cdot \texttt{(du, dv)} + \texttt{(u)(v)} \cdot \texttt{((du, dv))}</math>
+
|
 +
<math>\begin{matrix}
 +
uv \cdot \texttt{((} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{))} & + &
 +
u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)} & + &
 +
\texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)} & + &
 +
\texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{((} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{))}
 +
\end{matrix}</math>
 
|}
 
|}
  
<math>\operatorname{E}g</math> tells you what you would have to do, from where you are in the universe <math>[u, v],\!</math> if you want to end up in a place where <math>g\!</math> is true.  In this case, where the prevailing proposition <math>g\!</math> is <math>\texttt{((u, v))},</math> the component <math>\texttt{uv} \cdot \texttt{((du, dv))}</math> of <math>\operatorname{E}g</math> tells you this:  If <math>u\!</math> and <math>v\!</math> are both true where you are, then change either both or neither of <math>u\!</math> and <math>v\!</math> at the same time, and you will attain a place where <math>\texttt{((du, dv))}</math> is true.
+
In general, <math>\mathrm{E}g\!</math> tells you what you would have to do, from wherever you are in the universe <math>[u, v],\!</math> if you want to end up in a place where <math>g\!</math> is true.  In this case, where the prevailing proposition <math>g\!</math> is <math>\texttt{((} u \texttt{,~} v \texttt{))},\!</math> the component <math>uv \cdot \texttt{((} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{))}\!</math> of <math>\mathrm{E}g\!</math> tells you this:  If <math>u\!</math> and <math>v\!</math> are both true where you are, then change either both or neither of <math>u\!</math> and <math>v\!</math> at the same time, and you will attain a place where <math>\texttt{((} u \texttt{,~} v \texttt{))}\!</math> is true.
  
Figure&nbsp;2.3 shows the expansion of <math>\operatorname{D}g</math> over <math>[u, v]\!</math> to produce the expression:
+
Figure&nbsp;2.3 shows the expansion of <math>\mathrm{D}g\!</math> over <math>[u, v]\!</math> to produce the expression:
  
{| align="center" cellpadding="8" width="90%"
+
{| align="center" cellpadding="8" width="90%"
| <math>\texttt{uv} \cdot \texttt{(du, dv)} ~+~ \texttt{u(v)} \cdot \texttt{(du, dv)} ~+~ \texttt{(u)v} \cdot \texttt{(du, dv)} ~+~ \texttt{(u)(v)} \cdot \texttt{(du, dv)}</math>
+
|
 +
<math>\begin{matrix}
 +
uv \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)} & + &
 +
u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)} & + &
 +
\texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)} & + &
 +
\texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
\end{matrix}</math>
 
|}
 
|}
  
<math>\operatorname{D}g</math> tells you what you would have to do, from where you are in the universe <math>[u, v],\!</math> if you want to bring about a change in the value of <math>g,\!</math> that is, if you want to get to a place where the value of <math>g\!</math> is different from what it is where you are.  In the present case, where the ruling proposition <math>g\!</math> is <math>\texttt{((u, v))},</math> the term <math>\texttt{uv} \cdot \texttt{(du, dv)}</math> of <math>\operatorname{D}g</math> tells you this:  If <math>u\!</math> and <math>v\!</math> are both true where you are, then you would have to change one or the other but not both <math>u\!</math> and <math>v\!</math> in order to reach a place where the value of <math>g\!</math> is different from what it is where you are.
+
In general, <math>\mathrm{D}g\!</math> tells you what you would have to do, from wherever you are in the universe <math>[u, v],\!</math> if you want to bring about a change in the value of <math>g,\!</math> that is, if you want to get to a place where the value of <math>g\!</math> is different from what it is where you are.  In the present case, where the ruling proposition <math>g\!</math> is <math>\texttt{((} u \texttt{,~} v \texttt{))},\!</math> the term <math>uv \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}\!</math> of <math>\mathrm{D}g\!</math> tells you this:  If <math>u\!</math> and <math>v\!</math> are both true where you are, then you would have to change one or the other but not both <math>u\!</math> and <math>v\!</math> in order to reach a place where the value of <math>g\!</math> is different from what it is where you are.
  
Figure&nbsp;2.4 approximates <math>\operatorname{D}g</math> by the linear form <math>\operatorname{d}g</math> that expands over <math>[u, v]\!</math> as follows:
+
Figure&nbsp;2.4 approximates <math>\mathrm{D}g\!</math> by the linear form <math>{\mathrm{d}g}\!</math> that expands over <math>[u, v]\!</math> as follows:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{array}{lll}
+
<math>\begin{array}{*{9}{l}}
\operatorname{d}g
+
\mathrm{d}g
& = & \texttt{uv}\!\cdot\!\texttt{(du, dv)} + \texttt{u(v)}\!\cdot\!\texttt{(du, dv)} + \texttt{(u)v}\!\cdot\!\texttt{(du, dv)} + \texttt{(u)(v)}\!\cdot\!\texttt{(du, dv)}
+
& = & uv \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
\\ \\
+
& + & u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
& = & \texttt{(du, dv)}
+
& + & \texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 +
\\[8pt]
 +
& = & \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)}
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
Figure&nbsp;2.5 shows what remains of the difference map <math>\operatorname{D}g</math> when the first order linear contribution <math>\operatorname{d}g</math> is removed, namely:
+
Figure&nbsp;2.5 shows what remains of the difference map <math>\mathrm{D}g\!</math> when the first order linear contribution <math>{\mathrm{d}g}\!</math> is removed, namely:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
 
|
 
|
<math>\begin{matrix}
+
<math>\begin{array}{*{9}{l}}
\operatorname{r}g
+
\mathrm{r}g
& = & \texttt{uv} \cdot \texttt{0} & + & \texttt{u(v)} \cdot \texttt{0} & + & \texttt{(u)v} \cdot \texttt{0} & + & \texttt{(u)(v)} \cdot \texttt{0}
+
& = & uv \cdot 0
\\ \\
+
& + & u \texttt{(} v \texttt{)} \cdot 0
& = & \texttt{0}
+
& + & \texttt{(} u \texttt{)} v \cdot 0
\end{matrix}</math>
+
& + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot 0
 +
\\[8pt]
 +
& = & 0
 +
\end{array}</math>
 
|}
 
|}
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,550: Line 1,731:
 
Figure 2.1.  g = ((u, v))
 
Figure 2.1.  g = ((u, v))
 
</pre>
 
</pre>
 +
|}
  
<br>
+
{| align="center" border="0" cellpadding="10"
 
+
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,593: Line 1,775:
 
Figure 2.2.  Eg = ((u + du, v + dv))
 
Figure 2.2.  Eg = ((u + du, v + dv))
 
</pre>
 
</pre>
 +
|}
  
<br>
+
{| align="center" border="0" cellpadding="10"
 
+
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,636: Line 1,819:
 
Figure 2.3.  Difference Map Dg = g + Eg
 
Figure 2.3.  Difference Map Dg = g + Eg
 
</pre>
 
</pre>
 +
|}
  
<br>
+
{| align="center" border="0" cellpadding="10"
 
+
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,679: Line 1,863:
 
Figure 2.4.  Linear Proxy dg for Dg
 
Figure 2.4.  Linear Proxy dg for Dg
 
</pre>
 
</pre>
 +
|}
  
<br>
+
{| align="center" border="0" cellpadding="10"
 
+
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 1,722: Line 1,907:
 
Figure 2.5.  Remainder rg = Dg + dg
 
Figure 2.5.  Remainder rg = Dg + dg
 
</pre>
 
</pre>
 +
|}
  
 
<br>
 
<br>
Line 1,732: Line 1,918:
 
</pre>
 
</pre>
  
==Note 20==
+
==Visualization==
  
In my work on "[[Directory:Jon_Awbrey/Papers/Differential_Logic_and_Dynamic_Systems_2.0|Differential Logic and Dynamic Systems]]", I found it useful to develop several different ways of visualizing logical transformations, indeed, I devised four distinct styles of picture for the job.  Thus far in our work on the mapping <math>F : [u, v] \to [u, v],\!</math> we've been making use of what I call the ''areal view'' of the extended universe of discourse, <math>[u, v, du, dv],\!</math> but as the number of dimensions climbs beyond four, it's time to bid this genre adieu, and look for a style that can scale a little better.  At any rate, before we proceed any further, let's first assemble the information that we have gathered about <math>F\!</math> from several different angles, and see if it can be fitted into a coherent picture of the transformation <math>F : (u, v) \mapsto ( ~\texttt{((u)(v))}~, ~\texttt{((u, v))}~ ).</math>
+
In my work on [[Differential Logic and Dynamic Systems 2.0|Differential Logic and Dynamic Systems]], I found it useful to develop several different ways of visualizing logical transformations, indeed, I devised four distinct styles of picture for the job.  Thus far in our work on the mapping <math>F : [u, v] \to [u, v],\!</math> we've been making use of what I call the ''areal view'' of the extended universe of discourse, <math>[u, v, \mathrm{d}u, \mathrm{d}v],\!</math> but as the number of dimensions climbs beyond four, it's time to bid this genre adieu and look for a style that can scale a little better.  At any rate, before we proceed any further, let's first assemble the information that we have gathered about <math>F\!</math> from several different angles, and see if it can be fitted into a coherent picture of the transformation <math>F : (u, v) \mapsto ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ ).\!</math>
  
 
In our first crack at the transformation <math>F,\!</math> we simply plotted the state transitions and applied the utterly stock technique of calculating the finite differences.
 
In our first crack at the transformation <math>F,\!</math> we simply plotted the state transitions and applied the utterly stock technique of calculating the finite differences.
Line 1,743: Line 1,929:
 
|
 
|
 
<math>\begin{array}{c|cc|cc|}
 
<math>\begin{array}{c|cc|cc|}
t & u & v & du & dv \\[8pt]
+
t & u & v & \mathrm{d}u & \mathrm{d}v \\[8pt]
0 & 1 & 1 & 0 & 0 \\
+
0 & 1 & 1 & 0 & 0 \\
1 & '' & '' & '' & '' \\
+
1 & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
A quick inspection of the first Table suggests a rule to cover the case when <math>\texttt{u~=~v~=~1},</math> namely, <math>\texttt{du~=~dv~=~0}.</math>  To put it another way, the Table characterizes Orbit&nbsp;1 by means of the data:  <math>(u, v, du, dv) = (1, 1, 0, 0).\!</math>  Another way to convey the same information is by means of the extended proposition:  <math>\texttt{u~v~(du)(dv)}.</math>
+
A quick inspection of the first Table suggests a rule to cover the case when <math>u = v = 1,\!</math> namely, <math>\mathrm{d}u  = \mathrm{d}v = 0.\!</math>  To put it another way, the Table characterizes Orbit&nbsp;1 by means of the data:  <math>(u, v, \mathrm{d}u, \mathrm{d}v) = (1, 1, 0, 0).\!</math>  Another way to convey the same information is by means of the extended proposition:  <math>u v \texttt{(} \mathrm{d}u \texttt{)(} \mathrm{d}v \texttt{)}.\!</math>
  
 
{| align="center" cellpadding="8" style="text-align:center"
 
{| align="center" cellpadding="8" style="text-align:center"
Line 1,756: Line 1,942:
 
|
 
|
 
<math>\begin{array}{c|cc|cc|cc|}
 
<math>\begin{array}{c|cc|cc|cc|}
t & u & v & du & dv & d^2 u & d^2 v \\[8pt]
+
t & u & v & \mathrm{d}u & \mathrm{d}v & \mathrm{d}^2 u & \mathrm{d}^2 v \\[8pt]
0 & 0 & 0 & 0 & 1 &     1 &     0 \\
+
0 & 0 & 0 & 0 & 1 & 1 & 0 \\
1 & 0 & 1 & 1 & 1 &     1 &     1 \\
+
1 & 0 & 1 & 1 & 1 & 1 & 1 \\
2 & 1 & 0 & 0 & 0 &     0 &     0 \\
+
2 & 1 & 0 & 0 & 0 & 0 & 0 \\
3 & '' & '' & '' & '' &   '' &   '' \\
+
3 & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel
 
\end{array}</math>
 
\end{array}</math>
 
|}
 
|}
  
A more fine combing of the second Table brings to mind a rule that partly covers the remaining cases, that is, <math>\texttt{du~=~v}, ~\texttt{dv~=~(u)}.</math>  This much information about Orbit&nbsp;2 is also encapsulated by the extended proposition, <math>\texttt{(uv)((du, v))(dv, u)},</math> which says that <math>u\!</math> and <math>v\!</math> are not both true at the same time, while <math>du\!</math> is equal in value to <math>v\!</math> and <math>dv\!</math> is opposite in value to <math>u.\!</math>
+
A more fine combing of the second Table brings to mind a rule that partly covers the remaining cases, that is, <math>\mathrm{d}u = v, ~ \mathrm{d}v = \texttt{(} u \texttt{)}.\!</math>  This much information about Orbit&nbsp;2 is also encapsulated by the extended proposition <math>\texttt{(} uv \texttt{)((} \mathrm{d}u \texttt{,} v \texttt{))(} \mathrm{d}v, u \texttt{)},\!</math> which says that <math>u\!</math> and <math>v\!</math> are not both true at the same time, while <math>\mathrm{d}u\!</math> is equal in value to <math>v\!</math> and <math>\mathrm{d}v\!</math> is opposite in value to <math>u.\!</math>
 +
 
 +
==Turing Machine Example==
  
==Note 21==
+
<font size="3">&#9758;</font> See [[Theme One Program]] for documentation of the cactus graph syntax and the propositional modeling program used below.
  
By way of providing a simple illustration of Cook's Theorem, namely, that "Propositional Satisfiability is NP-Complete", I will describe one way to translate finite approximations of turing machines into propositional expressions, using the cactus language syntax for propositional calculus that I will describe in more detail as we proceed.
+
By way of providing a simple illustration of Cook's Theorem, namely, that &ldquo;Propositional Satisfiability is NP-Complete&rdquo;, I will describe one way to translate finite approximations of turing machines into propositional expressions, using the cactus language syntax for propositional calculus that I will describe in more detail as we proceed.
  
:; <math>\operatorname{Stilt}(k) =</math>
+
:; <math>\mathrm{Stilt}(k) =\!</math>
 
:: '''Space and time limited turing machine''',
 
:: '''Space and time limited turing machine''',
 
:: with <math>k\!</math> units of space and <math>k\!</math> units of time.
 
:: with <math>k\!</math> units of space and <math>k\!</math> units of time.
  
:; <math>\operatorname{Stunt}(k) =</math>
+
:; <math>\mathrm{Stunt}(k) =\!</math>
 
:: '''Space and time limited turing machine''',
 
:: '''Space and time limited turing machine''',
 
:: for computing the parity of a bit string, with number of tape cells of input equal to <math>k.\!</math>
 
:: for computing the parity of a bit string, with number of tape cells of input equal to <math>k.\!</math>
Line 1,782: Line 1,970:
 
A turing machine for computing the parity of a bit string is described by means of the following Figure and Table.
 
A turing machine for computing the parity of a bit string is described by means of the following Figure and Table.
  
<br>
+
{| align="center" border="0" cellspacing="10" style="text-align:center; width:100%"
 
+
| [[Image:Parity_Machine.jpg|400px]]
<pre>
+
|-
o-------------------------------------------------o
+
| height="20px" valign="top" | <math>\text{Figure 3.} ~~ \text{Parity Machine}\!</math>
|                                                 |
+
|}
|                    1/1/+1                      |
 
|                   -------->                    |
 
|                /\ /        \ /\                |
 
|       0/0/+1  ^  0          1  ^  0/0/+1      |
 
|                \/|\        /|\/                |
 
|                  | <-------- |                  |
 
|          #/#/-1  |  1/1/+1  |  #/#/-1          |
 
|                  |          |                  |
 
|                  v          v                  |
 
|                  #          *                  |
 
|                                                |
 
o-------------------------------------------------o
 
Figure 21-a. Parity Machine
 
</pre>
 
  
 
<br>
 
<br>
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
Table 21-b.  Parity Machine
+
Table 4.  Parity Machine
 
o-------o--------o-------------o---------o------------o
 
o-------o--------o-------------o---------o------------o
 
| State | Symbol | Next Symbol | Ratchet | Next State |
 
| State | Symbol | Next Symbol | Ratchet | Next State |
Line 1,818: Line 1,994:
 
o-------o--------o-------------o---------o------------o
 
o-------o--------o-------------o---------o------------o
 
</pre>
 
</pre>
 +
|}
  
 
<br>
 
<br>
  
The TM has a ''finite automaton'' (FA) as one component.  Let us refer to this particular FA by the name of <math>\operatorname{M}.</math>
+
The TM has a ''finite automaton'' (FA) as one component.  Let us refer to this particular FA by the name of <math>\mathrm{M}.</math>
  
The ''tape head'' (that is, the ''read unit'') will be called <math>\operatorname{H}.</math>  The ''registers'' are also called ''tape cells'' or ''tape squares''.
+
The ''tape head'' (that is, the ''read unit'') will be called <math>\mathrm{H}.</math>  The ''registers'' are also called ''tape cells'' or ''tape squares''.
  
==Note 22==
+
===Finite Approximations===
  
To see how each finite approximation to a given turing machine can be given a purely propositional description, one fixes the parameter <math>k\!</math> and limits the rest of the discussion to describing <math>\operatorname{Stilt}(k),</math> which is not really a full-fledged TM anymore but just a finite automaton in disguise.
+
To see how each finite approximation to a given turing machine can be given a purely propositional description, one fixes the parameter <math>k\!</math> and limits the rest of the discussion to describing <math>\mathrm{Stilt}(k),\!</math> which is not really a full-fledged TM anymore but just a finite automaton in disguise.
  
In this example, for the sake of a minimal illustration, we choose <math>k = 2,\!</math> and discuss <math>\operatorname{Stunt}(2).</math>  Since the zeroth tape cell and the last tape cell are both occupied by the character <math>^{\backprime\backprime}\texttt{\#}^{\prime\prime}</math> that is used for both the ''beginning of file'' <math>(\operatorname{bof})</math> and the ''end of file'' <math>(\operatorname{eof})</math> markers, this allows for only one digit of significant computation.
+
In this example, for the sake of a minimal illustration, we choose <math>k = 2,\!</math> and discuss <math>\mathrm{Stunt}(2).</math>  Since the zeroth tape cell and the last tape cell are both occupied by the character <math>^{\backprime\backprime}\texttt{\#}^{\prime\prime}</math> that is used for both the ''beginning of file'' <math>(\mathrm{bof})</math> and the ''end of file'' <math>(\mathrm{eof})</math> markers, this allows for only one digit of significant computation.
  
To translate <math>\operatorname{Stunt}(2)</math> into propositional form we use the following collection of basic propositions, boolean variables, or logical features, depending on what one prefers to call them:
+
To translate <math>\mathrm{Stunt}(2)</math> into propositional form we use the following collection of basic propositions, boolean variables, or logical features, depending on what one prefers to call them:
  
The basic propositions for describing the ''present state function'' <math>\operatorname{QF} : P \to Q</math> are these:
+
The basic propositions for describing the ''present state function'' <math>\mathrm{QF} : P \to Q</math> are these:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 1,851: Line 2,028:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
| At the point-in-time <math>p_i,\!</math> the finite state machine <math>\operatorname{M}</math> is in the state <math>q_j.\!</math>
+
| At the point-in-time <math>p_i,\!</math> the finite state machine <math>\mathrm{M}</math> is in the state <math>q_j.\!</math>
 
|}
 
|}
  
The basic propositions for describing the ''present register function'' <math>\operatorname{RF} : P \to R</math> are these:
+
The basic propositions for describing the ''present register function'' <math>\mathrm{RF} : P \to R</math> are these:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 1,872: Line 2,049:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
| At the point-in-time <math>p_i,\!</math> the tape head <math>\operatorname{H}</math> is on the tape cell <math>r_j.\!</math>
+
| At the point-in-time <math>p_i,\!</math> the tape head <math>\mathrm{H}</math> is on the tape cell <math>r_j.\!</math>
 
|}
 
|}
  
The basic propositions for describing the ''present symbol function'' <math>\operatorname{SF} : P \to (R \to S)</math> are these:
+
The basic propositions for describing the ''present symbol function'' <math>\mathrm{SF} : P \to (R \to S)</math> are these:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 1,887: Line 2,064:
 
\\[4pt]
 
\\[4pt]
 
\texttt{p0\_r3\_s\#}, & \texttt{p0\_r3\_s*}, & \texttt{p0\_r3\_s0}, & \texttt{p0\_r3\_s1},
 
\texttt{p0\_r3\_s\#}, & \texttt{p0\_r3\_s*}, & \texttt{p0\_r3\_s0}, & \texttt{p0\_r3\_s1},
\\[4pt]
+
\\[12pt]
 
\texttt{p1\_r0\_s\#}, & \texttt{p1\_r0\_s*}, & \texttt{p1\_r0\_s0}, & \texttt{p1\_r0\_s1},
 
\texttt{p1\_r0\_s\#}, & \texttt{p1\_r0\_s*}, & \texttt{p1\_r0\_s0}, & \texttt{p1\_r0\_s1},
 
\\[4pt]
 
\\[4pt]
Line 1,895: Line 2,072:
 
\\[4pt]
 
\\[4pt]
 
\texttt{p1\_r3\_s\#}, & \texttt{p1\_r3\_s*}, & \texttt{p1\_r3\_s0}, & \texttt{p1\_r3\_s1},
 
\texttt{p1\_r3\_s\#}, & \texttt{p1\_r3\_s*}, & \texttt{p1\_r3\_s0}, & \texttt{p1\_r3\_s1},
\\[4pt]
+
\\[12pt]
 
\texttt{p2\_r0\_s\#}, & \texttt{p2\_r0\_s*}, & \texttt{p2\_r0\_s0}, & \texttt{p2\_r0\_s1},
 
\texttt{p2\_r0\_s\#}, & \texttt{p2\_r0\_s*}, & \texttt{p2\_r0\_s0}, & \texttt{p2\_r0\_s1},
 
\\[4pt]
 
\\[4pt]
Line 1,903: Line 2,080:
 
\\[4pt]
 
\\[4pt]
 
\texttt{p2\_r3\_s\#}, & \texttt{p2\_r3\_s*}, & \texttt{p2\_r3\_s0}, & \texttt{p2\_r3\_s1},
 
\texttt{p2\_r3\_s\#}, & \texttt{p2\_r3\_s*}, & \texttt{p2\_r3\_s0}, & \texttt{p2\_r3\_s1},
\\[4pt]
+
\\[12pt]
 
\texttt{p3\_r0\_s\#}, & \texttt{p3\_r0\_s*}, & \texttt{p3\_r0\_s0}, & \texttt{p3\_r0\_s1},
 
\texttt{p3\_r0\_s\#}, & \texttt{p3\_r0\_s*}, & \texttt{p3\_r0\_s0}, & \texttt{p3\_r0\_s1},
 
\\[4pt]
 
\\[4pt]
Line 1,910: Line 2,087:
 
\texttt{p3\_r2\_s\#}, & \texttt{p3\_r2\_s*}, & \texttt{p3\_r2\_s0}, & \texttt{p3\_r2\_s1},
 
\texttt{p3\_r2\_s\#}, & \texttt{p3\_r2\_s*}, & \texttt{p3\_r2\_s0}, & \texttt{p3\_r2\_s1},
 
\\[4pt]
 
\\[4pt]
\texttt{p3\_r3\_s\#}, & \texttt{p3\_r3\_s*}, & \texttt{p3\_r3\_s0}, & \texttt{p3\_r3\_s1},
+
\texttt{p3\_r3\_s\#}, & \texttt{p3\_r3\_s*}, & \texttt{p3\_r3\_s0}, & \texttt{p3\_r3\_s1}.
\\[4pt]
 
 
\end{matrix}</math>
 
\end{matrix}</math>
 
|}
 
|}
Line 1,921: Line 2,097:
 
|}
 
|}
  
==Initial Conditions==
+
===Initial Conditions===
  
Given but a single free square on the tape, there are just two different sets of initial conditions for <math>\operatorname{Stunt}(2),</math> the finite approximation to the parity turing machine that we are presently considering.
+
Given but a single free square on the tape, there are just two different sets of initial conditions for <math>\mathrm{Stunt}(2),</math> the finite approximation to the parity turing machine that we are presently considering.
  
===Initial Conditions for Tape Input "0"===
+
====Initial Conditions for Tape Input "0"====
  
The following conjunction of 5 basic propositions describes the initial conditions when <math>\operatorname{Stunt}(2)</math> is started with an input of "0" in its free square:
+
The following conjunction of 5 basic propositions describes the initial conditions when <math>\mathrm{Stunt}(2)</math> is started with an input of "0" in its free square:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 1,948: Line 2,124:
 
{| align="center" cellpadding=8" width="90%"
 
{| align="center" cellpadding=8" width="90%"
 
|
 
|
<p>At time <math>p_0,\!</math> machine <math>\operatorname{M}</math> is in the state <math>q_0,\!</math></p>
+
<p>At time <math>p_0,\!</math> machine <math>\mathrm{M}</math> is in the state <math>q_0,\!</math></p>
<p>At time <math>p_0,\!</math> scanner <math>\operatorname{H}</math> is reading cell <math>r_1,\!</math></p>
+
<p>At time <math>p_0,\!</math> scanner <math>\mathrm{H}</math> is reading cell <math>r_1,\!</math></p>
 
<p>At time <math>p_0,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math></p>
 
<p>At time <math>p_0,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math></p>
 
<p>At time <math>p_0,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{0},</math></p>
 
<p>At time <math>p_0,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{0},</math></p>
Line 1,955: Line 2,131:
 
|}
 
|}
  
===Initial Conditions for Tape Input "1"===
+
====Initial Conditions for Tape Input "1"====
  
The following conjunction of 5 basic propositions describes the initial conditions when <math>\operatorname{Stunt}(2)</math> is started with an input of "1" in its free square:
+
The following conjunction of 5 basic propositions describes the initial conditions when <math>\mathrm{Stunt}(2)</math> is started with an input of "1" in its free square:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 1,978: Line 2,154:
 
{| align="center" cellpadding=8" width="90%"
 
{| align="center" cellpadding=8" width="90%"
 
|
 
|
<p>At time <math>p_0,\!</math> machine <math>\operatorname{M}</math> is in the state <math>q_0,\!</math></p>
+
<p>At time <math>p_0,\!</math> machine <math>\mathrm{M}</math> is in the state <math>q_0,\!</math></p>
<p>At time <math>p_0,\!</math> scanner <math>\operatorname{H}</math> is reading cell <math>r_1,\!</math></p>
+
<p>At time <math>p_0,\!</math> scanner <math>\mathrm{H}</math> is reading cell <math>r_1,\!</math></p>
 
<p>At time <math>p_0,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math></p>
 
<p>At time <math>p_0,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math></p>
 
<p>At time <math>p_0,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{1},</math></p>
 
<p>At time <math>p_0,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{1},</math></p>
Line 1,985: Line 2,161:
 
|}
 
|}
  
==Propositional Program==
+
===Propositional Program===
  
A complete description of <math>\operatorname{Stunt}(2)</math> in propositional form is obtained by conjoining one of the above choices for initial conditions with all of the following sets of propositions, that serve in effect as a simple type of ''declarative program'', telling us all that we need to know about the anatomy and behavior of the truncated TM in question.
+
A complete description of <math>\mathrm{Stunt}(2)</math> in propositional form is obtained by conjoining one of the above choices for initial conditions with all of the following sets of propositions, that serve in effect as a simple type of ''declarative program'', telling us all that we need to know about the anatomy and behavior of the truncated TM in question.
  
===Mediate Conditions===
+
====Mediate Conditions====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,004: Line 2,180:
 
|}
 
|}
  
===Terminal Conditions===
+
====Terminal Conditions====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,013: Line 2,189:
 
|}
 
|}
  
===State Partition===
+
====State Partition====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,026: Line 2,202:
 
|}
 
|}
  
===Register Partition===
+
====Register Partition====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,039: Line 2,215:
 
|}
 
|}
  
===Symbol Partition===
+
====Symbol Partition====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,064: Line 2,240:
 
|}
 
|}
  
===Interaction Conditions===
+
====Interaction Conditions====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,107: Line 2,283:
 
|}
 
|}
  
===Transition Relations===
+
====Transition Relations====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,146: Line 2,322:
 
|}
 
|}
  
==Interpretation of the Propositional Program==
+
===Interpretation of the Propositional Program===
  
Let us now run through the propositional specification of <math>\operatorname{Stunt}(2),</math> our truncated TM, and paraphrase what it says in ordinary language.
+
Let us now run through the propositional specification of <math>\mathrm{Stunt}(2),</math> our truncated TM, and paraphrase what it says in ordinary language.
  
===Mediate Conditions===
+
====Mediate Conditions====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,170: Line 2,346:
 
|
 
|
 
<math>\begin{array}{l}
 
<math>\begin{array}{l}
\operatorname{not}~ p ~\operatorname{without}~ q
+
\mathrm{not}~ p ~\mathrm{without}~ q
 
\\[4pt]
 
\\[4pt]
p ~\operatorname{implies}~ q
+
p ~\mathrm{implies}~ q
 
\\[4pt]
 
\\[4pt]
\operatorname{if}~ p ~\operatorname{then}~ q
+
\mathrm{if}~ p ~\mathrm{then}~ q
 
\\[4pt]
 
\\[4pt]
 
p \Rightarrow q
 
p \Rightarrow q
Line 2,184: Line 2,360:
 
<br>
 
<br>
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 2,196: Line 2,374:
 
o---------------------------------------o
 
o---------------------------------------o
 
</pre>
 
</pre>
 +
|}
  
 
<br>
 
<br>
Line 2,203: Line 2,382:
 
{| align="center" cellpadding=8" width="90%"
 
{| align="center" cellpadding=8" width="90%"
 
|
 
|
<p>If <math>\operatorname{M}</math> at <math>p_0\!</math> is in state <math>q_\#,\!</math> then <math>\operatorname{M}</math> at <math>p_1\!</math> is in state <math>q_\#,\!</math> and</p>
+
<p>If <math>\mathrm{M}</math> at <math>p_0\!</math> is in state <math>q_\#,\!</math> then <math>\mathrm{M}</math> at <math>p_1\!</math> is in state <math>q_\#,\!</math> and</p>
  
<p>If <math>\operatorname{M}</math> at <math>p_0\!</math> is in state <math>q_*,\!</math> then <math>\operatorname{M}</math> at <math>p_1\!</math> is in state <math>q_*,\!</math> and</p>
+
<p>If <math>\mathrm{M}</math> at <math>p_0\!</math> is in state <math>q_*,\!</math> then <math>\mathrm{M}</math> at <math>p_1\!</math> is in state <math>q_*,\!</math> and</p>
  
<p>If <math>\operatorname{M}</math> at <math>p_1\!</math> is in state <math>q_\#,\!</math> then <math>\operatorname{M}</math> at <math>p_2\!</math> is in state <math>q_\#,\!</math> and</p>
+
<p>If <math>\mathrm{M}</math> at <math>p_1\!</math> is in state <math>q_\#,\!</math> then <math>\mathrm{M}</math> at <math>p_2\!</math> is in state <math>q_\#,\!</math> and</p>
  
<p>If <math>\operatorname{M}</math> at <math>p_1\!</math> is in state <math>q_*,\!</math> then <math>\operatorname{M}</math> at <math>p_2\!</math> is in state <math>q_*.\!</math></p>
+
<p>If <math>\mathrm{M}</math> at <math>p_1\!</math> is in state <math>q_*,\!</math> then <math>\mathrm{M}</math> at <math>p_2\!</math> is in state <math>q_*.\!</math></p>
 
|}
 
|}
  
===Terminal Conditions===
+
====Terminal Conditions====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,221: Line 2,400:
 
|}
 
|}
  
In cactus syntax, an expression of the form <math>\texttt{((p)(q))}</math> expresses the disjunction <math>p ~\operatorname{or}~ q.</math>  The corresponding cactus graph, here just a tree, has the following shape:
+
In cactus syntax, an expression of the form <math>\texttt{((p)(q))}</math> expresses the disjunction <math>p ~\mathrm{or}~ q.</math>  The corresponding cactus graph, here just a tree, has the following shape:
  
 
<br>
 
<br>
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 2,239: Line 2,420:
 
o---------------------------------------o
 
o---------------------------------------o
 
</pre>
 
</pre>
 +
|}
  
 
<br>
 
<br>
Line 2,246: Line 2,428:
 
{| align="center" cellpadding=8" width="90%"
 
{| align="center" cellpadding=8" width="90%"
 
|
 
|
<p>At time <math>p_2\!</math> machine <math>\operatorname{M}</math> is in state <math>q_\#,\!</math> or</p>
+
<p>At time <math>p_2\!</math> machine <math>\mathrm{M}</math> is in state <math>q_\#,\!</math> or</p>
<p>At time <math>p_2\!</math> machine <math>\operatorname{M}</math> is in state <math>q_*.\!</math></p>
+
<p>At time <math>p_2\!</math> machine <math>\mathrm{M}</math> is in state <math>q_*.\!</math></p>
 
|}
 
|}
  
===State Partition===
+
====State Partition====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,263: Line 2,445:
 
|}
 
|}
  
In cactus syntax, an expression of the form <math>\texttt{((} e_1 \texttt{),(} e_2 \texttt{),(} \ldots \texttt{),(} e_k \texttt{))}</math> expresses a statement to the effect that ''exactly one of the expressions <math>e_j\!</math> is true, for <math>j = 1 ~\mathit{to}~ k.</math>''  Expressions of this form are called ''universal partition'' expressions, and the corresponding ''painted and rooted cactus'' (PARC) has the following shape:
+
In cactus syntax, an expression of the form <math>\texttt{((} e_1 \texttt{),(} e_2 \texttt{),(} \ldots \texttt{),(} e_k \texttt{))}\!</math> expresses a statement to the effect that ''exactly one of the expressions <math>e_j\!</math> is true, for <math>j = 1 ~\mathit{to}~ k.</math>''  Expressions of this form are called ''universal partition'' expressions, and the corresponding ''painted and rooted cactus'' (PARC) has the following shape:
  
 
<br>
 
<br>
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 2,288: Line 2,472:
 
o---------------------------------------o
 
o---------------------------------------o
 
</pre>
 
</pre>
 +
|}
  
 
<br>
 
<br>
  
The State Partition segment of the propositional program consists of three universal partition expressions, taken in conjunction expressing the condition that <math>\operatorname{M}</math> has to be in one and only one of its states at each point in time under consideration.  In short, we have the constraint:
+
The State Partition segment of the propositional program consists of three universal partition expressions, taken in conjunction expressing the condition that <math>\mathrm{M}</math> has to be in one and only one of its states at each point in time under consideration.  In short, we have the constraint:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,297: Line 2,482:
 
<p>At each of the points in time <math>p_i,\!</math> for <math>i\!</math> in the set <math>\{ 0, 1, 2 \},\!</math></p>
 
<p>At each of the points in time <math>p_i,\!</math> for <math>i\!</math> in the set <math>\{ 0, 1, 2 \},\!</math></p>
  
<p><math>\operatorname{M}</math> can be in exactly one state <math>q_j,\!</math> for <math>j\!</math> in the set <math>\{ 0, 1, \#, * \}.</math></p>
+
<p><math>\mathrm{M}</math> can be in exactly one state <math>q_j,\!</math> for <math>j\!</math> in the set <math>\{ 0, 1, \#, * \}.\!</math></p>
 
|}
 
|}
  
===Register Partition===
+
====Register Partition====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,313: Line 2,498:
 
|}
 
|}
  
The Register Partition segment of the propositional program consists of three universal partition expressions, taken in conjunction saying that the read head <math>H\!</math> must be reading one and only one of the registers or tape cells available to it at each of the points in time under consideration.  In sum:
+
The Register Partition segment of the propositional program consists of three universal partition expressions, taken in conjunction saying that the read head <math>\mathrm{H}</math> must be reading one and only one of the registers or tape cells available to it at each of the points in time under consideration.  In sum:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,319: Line 2,504:
 
<p>At each of the points in time <math>p_i,\!</math> for <math>i = 0, 1, 2,\!</math></p>
 
<p>At each of the points in time <math>p_i,\!</math> for <math>i = 0, 1, 2,\!</math></p>
  
<p><math>H\!</math> is reading exactly one cell <math>r_j,\!</math> for <math>j = 0, 1, 2.\!</math>
+
<p><math>\mathrm{H}</math> is reading exactly one cell <math>r_j,\!</math> for <math>j = 0, 1, 2.\!</math>
 
|}
 
|}
  
===Symbol Partition===
+
====Symbol Partition====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,347: Line 2,532:
 
|}
 
|}
  
The Symbol Partition segment of the propositional program for <math>\operatorname{Stunt}(2)</math> consists of nine universal partition expressions, taken in conjunction stipulating that there has to be one and only one symbol in each of the registers at each point in time under consideration.  In short, we have:
+
The Symbol Partition segment of the propositional program for <math>\mathrm{Stunt}(2)</math> consists of nine universal partition expressions, taken in conjunction stipulating that there has to be one and only one symbol in each of the registers at each point in time under consideration.  In short, we have:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,358: Line 2,543:
 
|}
 
|}
  
===Interaction Conditions===
+
====Interaction Conditions====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,414: Line 2,599:
 
This expression has the corresponding cactus graph:
 
This expression has the corresponding cactus graph:
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 
<pre>
 
<pre>
 
o---------------------------------------o
 
o---------------------------------------o
Line 2,426: Line 2,613:
 
o---------------------------------------o
 
o---------------------------------------o
 
</pre>
 
</pre>
 +
|}
  
 
A propositional expression of this form can be read as follows:
 
A propositional expression of this form can be read as follows:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
|<math>\operatorname{If}</math>
+
|<math>\mathrm{If}</math>
 
|-
 
|-
 
| At the time <math>p_i,\!</math> the tape cell <math>r_j\!</math> bears the mark <math>s_k,\!</math>
 
| At the time <math>p_i,\!</math> the tape cell <math>r_j\!</math> bears the mark <math>s_k,\!</math>
 
|-
 
|-
| <math>\operatorname{But}</math> it is not the case that:
+
| <math>\mathrm{But}</math> it is not the case that:
 
|-
 
|-
 
| At the time <math>p_i,\!</math> the tape head is on the tape cell <math>r_j,\!</math>
 
| At the time <math>p_i,\!</math> the tape head is on the tape cell <math>r_j,\!</math>
 
|-
 
|-
| <math>\operatorname{Then}</math>
+
| <math>\mathrm{Then}</math>
 
|-
 
|-
 
| At the time <math>p_{i+1},\!</math> the tape cell <math>r_j\!</math> bears the mark <math>s_k.\!</math>
 
| At the time <math>p_{i+1},\!</math> the tape cell <math>r_j\!</math> bears the mark <math>s_k.\!</math>
Line 2,445: Line 2,633:
 
The eighteen clauses of the Interaction Conditions simply impose one such constraint on symbol changes for each combination of the times <math>p_0, p_1,\!</math> registers <math>r_0, r_1, r_2,\!</math> and symbols <math>s_0, s_1, s_\#.\!</math>
 
The eighteen clauses of the Interaction Conditions simply impose one such constraint on symbol changes for each combination of the times <math>p_0, p_1,\!</math> registers <math>r_0, r_1, r_2,\!</math> and symbols <math>s_0, s_1, s_\#.\!</math>
  
===Transition Relations===
+
====Transition Relations====
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
Line 2,484: Line 2,672:
 
|}
 
|}
  
The Transition Relation segment of the propositional program for <math>\operatorname{Stunt}(2)</math> consists of sixteen implication statements with complex antecedents and consequents.  Taken together, these give propositional expression to the TM Figure and Table that were given at the outset.
+
The Transition Relation segment of the propositional program for <math>\mathrm{Stunt}(2)</math> consists of sixteen implication statements with complex antecedents and consequents.  Taken together, these give propositional expression to the TM Figure and Table that were given at the outset.
  
 
Just by way of a single example, consider the clause:
 
Just by way of a single example, consider the clause:
Line 2,495: Line 2,683:
  
 
{| align="center" cellpadding="8" width="90%"
 
{| align="center" cellpadding="8" width="90%"
|<math>\operatorname{If}</math>
+
|<math>\mathrm{If}</math>
 
|-
 
|-
| At the time <math>p_0,\!</math> the machine <math>M\!</math> is in the state <math>q_0,\!</math> and
+
| At the time <math>p_0,\!</math> the machine <math>\mathrm{M}</math> is in the state <math>q_0,\!</math> and
 
|-
 
|-
| At the time <math>p_0,\!</math> the scanner <math>H\!</math> is reading cell <math>r_1,\!</math> and
+
| At the time <math>p_0,\!</math> the scanner <math>\mathrm{H}</math> is reading cell <math>r_1,\!</math> and
 
|-
 
|-
 
| At the time <math>p_0,\!</math> the tape cell <math>r_1\!</math> contains a <math>\texttt{1},</math>
 
| At the time <math>p_0,\!</math> the tape cell <math>r_1\!</math> contains a <math>\texttt{1},</math>
 
|-
 
|-
| <math>\operatorname{Then}</math>
+
| <math>\mathrm{Then}</math>
 
|-
 
|-
| At the time <math>p_1,\!</math> the machine <math>M\!</math> is in the state <math>q_1,\!</math> and
+
| At the time <math>p_1,\!</math> the machine <math>\mathrm{M}</math> is in the state <math>q_1,\!</math> and
 
|-
 
|-
| At the time <math>p_1,\!</math> the scanner <math>H\!</math> is reading cell <math>r_2,\!</math> and
+
| At the time <math>p_1,\!</math> the scanner <math>\mathrm{H}</math> is reading cell <math>r_2,\!</math> and
 
|-
 
|-
| At the time <math>p_1,\!</math> the tape cell <math>r_1\!</math> contains a <math>\texttt{1},</math>
+
| At the time <math>p_1,\!</math> the tape cell <math>r_1\!</math> contains a <math>\texttt{1}.</math>
 
|}
 
|}
  
==Note 32==
+
===Computation===
  
<pre>
+
The propositional program for <math>\mathrm{Stunt}(2)</math> uses the following set
Interpretation of the Propositional Program (cont.)
+
of <math>9 + 12 + 36 = 57\!</math> basic propositions or boolean variables:
  
The propositional program for Stunt(2) uses the following set
+
{| align="center" cellpadding="8" width="90%"
of (9 + 12 + 36) = 57 basic propositions or boolean variables:
+
|
 +
<math>\begin{matrix}
 +
\texttt{p0\_r0}, & \texttt{p0\_r1}, & \texttt{p0\_r2},
 +
\\[6pt]
 +
\texttt{p1\_r0}, & \texttt{p1\_r1}, & \texttt{p1\_r2},
 +
\\[6pt]
 +
\texttt{p2\_r0}, & \texttt{p2\_r1}, & \texttt{p2\_r2}.
 +
\end{matrix}</math>
 +
|}
  
  p0_r0, p0_r1, p0_r2,
+
{| align="center" cellpadding="8" width="90%"
  p1_r0, p1_r1, p1_r2,
+
|
  p2_r0, p2_r1, p2_r2.
+
<math>\begin{matrix}
 +
\texttt{p0\_q\#}, & \texttt{p0\_q*}, & \texttt{p0\_q0}, & \texttt{p0\_q1},
 +
\\[6pt]
 +
\texttt{p1\_q\#}, & \texttt{p1\_q*}, & \texttt{p1\_q0}, & \texttt{p1\_q1},
 +
\\[6pt]
 +
\texttt{p2\_q\#}, & \texttt{p2\_q*}, & \texttt{p2\_q0}, & \texttt{p2\_q1}.
 +
\end{matrix}</math>
 +
|}
  
  p0_q#, p0_q*, p0_q0, p0_q1,
+
{| align="center" cellpadding="8" width="90%"
  p1_q#, p1_q*, p1_q0, p1_q1,
+
|
  p2_q#, p2_q*, p2_q0, p2_q1.
+
<math>\begin{matrix}
 
+
\texttt{p0\_r0\_s\#}, & \texttt{p0\_r0\_s*}, & \texttt{p0\_r0\_s0}, & \texttt{p0\_r0\_s1},
  p0_r0_s#, p0_r0_s*, p0_r0_s0, p0_r0_s1,
+
\\[4pt]
  p0_r1_s#, p0_r1_s*, p0_r1_s0, p0_r1_s1,
+
\texttt{p0\_r1\_s\#}, & \texttt{p0\_r1\_s*}, & \texttt{p0\_r1\_s0}, & \texttt{p0\_r1\_s1},
  p0_r2_s#, p0_r2_s*, p0_r2_s0, p0_r2_s1,
+
\\[4pt]
 
+
\texttt{p0\_r2\_s\#}, & \texttt{p0\_r2\_s*}, & \texttt{p0\_r2\_s0}, & \texttt{p0\_r2\_s1},
  p1_r0_s#, p1_r0_s*, p1_r0_s0, p1_r0_s1,
+
\\[12pt]
  p1_r1_s#, p1_r1_s*, p1_r1_s0, p1_r1_s1,
+
\texttt{p1\_r0\_s\#}, & \texttt{p1\_r0\_s*}, & \texttt{p1\_r0\_s0}, & \texttt{p1\_r0\_s1},
  p1_r2_s#, p1_r2_s*, p1_r2_s0, p1_r2_s1,
+
\\[4pt]
 
+
\texttt{p1\_r1\_s\#}, & \texttt{p1\_r1\_s*}, & \texttt{p1\_r1\_s0}, & \texttt{p1\_r1\_s1},
  p2_r0_s#, p2_r0_s*, p2_r0_s0, p2_r0_s1,
+
\\[4pt]
  p2_r1_s#, p2_r1_s*, p2_r1_s0, p2_r1_s1,
+
\texttt{p1\_r2\_s\#}, & \texttt{p1\_r2\_s*}, & \texttt{p1\_r2\_s0}, & \texttt{p1\_r2\_s1},
  p2_r2_s#, p2_r2_s*, p2_r2_s0, p2_r2_s1.
+
\\[12pt]
 
+
\texttt{p2\_r0\_s\#}, & \texttt{p2\_r0\_s*}, & \texttt{p2\_r0\_s0}, & \texttt{p2\_r0\_s1},
This means that the propositional program itself is nothing more or
+
\\[4pt]
less than a single proposition or a boolean function P : B^57 -> B.
+
\texttt{p2\_r1\_s\#}, & \texttt{p2\_r1\_s*}, & \texttt{p2\_r1\_s0}, & \texttt{p2\_r1\_s1},
 
+
\\[4pt]
An assignment of boolean values to the above set of boolean variables
+
\texttt{p2\_r2\_s\#}, & \texttt{p2\_r2\_s*}, & \texttt{p2\_r2\_s0}, & \texttt{p2\_r2\_s1}.
is called an "interpretation" of P, and any interpretation of P that
+
\end{matrix}</math>
makes the proposition P : B^57 -> B evaluate to 1 is referred to as
+
|}
a "satisfying interpretation" of the proposition P.  Another way to
+
 
specify interpretations, instead of giving them as bit vectors in
+
This means that the propositional program itself is nothing but a single proposition or boolean function of the form <math>p : \mathbb{B}^{57} \to \mathbb{B}.</math>
B^57 and trying to remember some arbitrary ordering of variables,
+
 
is to give them in the form of "singular propositions", that is,
+
An assignment of boolean values to the above set of boolean variables is called an ''interpretation'' of the proposition <math>p,\!</math> and any interpretation of <math>p\!</math> that makes the proposition <math>p : \mathbb{B}^{57} \to \mathbb{B}</math> evaluate to <math>1\!</math> is referred to as a ''satisfying interpretation'' of the proposition <math>p.\!</math> Another way to specify interpretations, instead of giving them as bit vectors in <math>\mathbb{B}^{57}</math> and trying to remember some arbitrary ordering of variables, is to give them in the form of ''singular propositions'', that is, a conjunction of the form <math>e_1 \cdot \ldots \cdot e_{57}</math> where each <math>e_j\!</math> is either <math>v_j\!</math> or <math>\texttt{(} v_j \texttt{)},</math> that is, either the assertion or the negation of the boolean variable <math>{v_j},\!</math> as <math>j\!</math> runs from 1 to 57. Even more briefly, the same information can be communicated simply by giving the conjunction of the asserted variables, with the understanding that each of the others is negated.
a conjunction of the form "e_1 & ... & e_57" where each e_j is
+
 
either "v_j" or "(v_j)", that is, either the assertion or the
+
A satisfying interpretation of the proposition <math>p\!</math> supplies us with all the information of a complete execution history for the corresponding program, and so all we have to do in order to get the output of the program <math>p\!</math> is to read off the proper part of the data from the expression of this interpretation.
negation of the boolean variable v_j, as j runs from 1 to 57.
 
Even more briefly, the same information can be communicated
 
simply by giving the conjunction of the asserted variables,
 
with the understanding that each of the others is negated.
 
 
 
A satisfying interpretation of the proposition P supplies us
 
with all the information of a complete execution history for
 
the corresponding program, and so all we have to do in order
 
to get the output of the program P is to read off the proper
 
part of the data from the expression of this interpretation.
 
</pre>
 
  
==Note 33==
+
===Output===
  
<pre>
+
One component of the <math>\begin{smallmatrix}\mathrm{Theme~One}\end{smallmatrix}</math> program that I wrote some years ago finds all the satisfying interpretations of propositions expressed in cactus syntax.  It's not a polynomial time algorithm, as you may guess, but it was just barely efficient enough to do this example in the 500 K of spare memory that I had on an old 286 PC in about 1989, so I will give you the actual outputs from those trials.
Interpretation of the Propositional Program (concl.)
 
  
One component of the Theme One program that I wrote some years ago
+
====Output Conditions for Tape Input "0"====
finds all the satisfying interpretations of propositions expressed
 
in cactus syntax.  It's not a polynomial time algorithm, as you may
 
guess, but it was just barely efficient enough to do this example
 
in the 500 K of spare memory that I had on an old 286 PC in about
 
1989, so I will give you the actual outputs from those trials.
 
  
Output Conditions for Tape Input "0"
+
Let <math>p_0\!</math> be the proposition that we get by conjoining the proposition that describes the initial conditions for tape input "0" with the proposition that describes the truncated turing machine <math>\mathrm{Stunt}(2).</math>  As it turns out, <math>p_0\!</math> has a single satisfying interpretation.  This interpretation is expressible in the form of a singular proposition, which can in turn be indicated by its positive logical features, as shown in the following display:
  
Let P_0 be the proposition that we get by conjoining
+
<br>
the proposition that describes the initial conditions
 
for tape input "0" with the proposition that describes
 
the truncated turing machine Stunt(2).  As it turns out,
 
P_0 has a single satisfying interpretation, and this is
 
represented as a singular proposition in terms of its
 
positive logical features in the following display:
 
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 +
<pre>
 
o-------------------------------------------------o
 
o-------------------------------------------------o
 
|                                                |
 
|                                                |
Line 2,605: Line 2,788:
 
|                                                |
 
|                                                |
 
o-------------------------------------------------o
 
o-------------------------------------------------o
 +
</pre>
 +
|}
 +
 +
<br>
  
 
The Output Conditions for Tape Input "0" can be read as follows:
 
The Output Conditions for Tape Input "0" can be read as follows:
  
  At the time p_0, M is in the state q_0, and
+
{| align="center" cellpadding=8" width="90%"
  At the time p_0, H is reading cell r_1, and
+
|
  At the time p_0, cell r_0 contains "#", and
+
<p>At the time <math>p_0,\!</math> machine <math>\mathrm{M}</math> is in the state <math>q_0,\!</math> and</p>
  At the time p_0, cell r_1 contains "0", and
+
<p>At the time <math>p_0,\!</math> scanner <math>\mathrm{H}</math> is reading cell <math>r_1,\!</math> and</p>
  At the time p_0, cell r_2 contains "#", and
+
<p>At the time <math>p_0,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
<p>At the time <math>p_0,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{0},</math> and</p>
 +
<p>At the time <math>p_0,\!</math> cell <math>r_2\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
|-
 +
|
 +
<p>At the time <math>p_1,\!</math> machine <math>\mathrm{M}</math> is in the state <math>q_0,\!</math> and</p>
 +
<p>At the time <math>p_1,\!</math> scanner <math>\mathrm{H}</math> is reading cell <math>r_2,\!</math> and</p>
 +
<p>At the time <math>p_1,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
<p>At the time <math>p_1,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{0},</math> and</p>
 +
<p>At the time <math>p_1,\!</math> cell <math>r_2\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
|-
 +
|
 +
<p>At the time <math>p_2,\!</math> machine <math>\mathrm{M}</math> is in the state <math>q_\#,\!</math> and</p>
 +
<p>At the time <math>p_2,\!</math> scanner <math>\mathrm{H}</math> is reading cell <math>r_1,\!</math> and</p>
 +
<p>At the time <math>p_2,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
<p>At the time <math>p_2,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{0},</math> and</p>
 +
<p>At the time <math>p_2,\!</math> cell <math>r_2\!</math> contains the symbol <math>\texttt{\#}.</math></p>
 +
|}
  
  At the time p_1, M is in the state q_0, and
+
The output of <math>\mathrm{Stunt}(2)</math> being the symbol that rests under the tape head <math>\mathrm{H}</math> if and when the machine <math>\mathrm{M}</math> reaches one of its resting states, we get the result that <math>\mathrm{Parity}(0) = 0.</math>
  At the time p_1, H is reading cell r_2, and
 
  At the time p_1, cell r_0 contains "#", and
 
  At the time p_1, cell r_1 contains "0", and
 
  At the time p_1, cell r_2 contains "#", and
 
  
  At the time p_2, M is in the state q_#, and
+
====Output Conditions for Tape Input "1"====
  At the time p_2, H is reading cell r_1, and
 
  At the time p_2, cell r_0 contains "#", and
 
  At the time p_2, cell r_1 contains "0", and
 
  At the time p_2, cell r_2 contains "#".
 
  
The output of Stunt(2) being the symbol that rests under
+
Let <math>p_1\!</math> be the proposition that we get by conjoining the proposition that describes the initial conditions for tape input "1" with the proposition that describes the truncated turing machine <math>\mathrm{Stunt}(2).</math>  As it turns out, <math>p_1\!</math> has a single satisfying interpretation.  This interpretation is expressible in the form of a singular proposition, which can in turn be indicated by its positive logical features, as shown in the following display:
the tape head H if and when the machine M reaches one of
 
its resting states, we get the result that Parity(0) = 0.
 
  
Output Conditions for Tape Input "1"
+
<br>
 
 
Let P_1 be the proposition that we get by conjoining
 
the proposition that describes the initial conditions
 
for tape input "1" with the proposition that describes
 
the truncated turing machine Stunt(2).  As it turns out,
 
P_1 has a single satisfying interpretation, and this is
 
represented as a singular proposition in terms of its
 
positive logical features in the following display:
 
  
 +
{| align="center" border="0" cellpadding="10"
 +
|
 +
<pre>
 
o-------------------------------------------------o
 
o-------------------------------------------------o
 
|                                                |
 
|                                                |
Line 2,659: Line 2,848:
 
|                                                |
 
|                                                |
 
o-------------------------------------------------o
 
o-------------------------------------------------o
 +
</pre>
 +
|}
 +
 +
<br>
  
 
The Output Conditions for Tape Input "1" can be read as follows:
 
The Output Conditions for Tape Input "1" can be read as follows:
  
  At the time p_0, M is in the state q_0, and
+
{| align="center" cellpadding=8" width="90%"
  At the time p_0, H is reading cell r_1, and
+
|
  At the time p_0, cell r_0 contains "#", and
+
<p>At the time <math>p_0,\!</math> machine <math>\mathrm{M}</math> is in the state <math>q_0,\!</math> and</p>
  At the time p_0, cell r_1 contains "1", and
+
<p>At the time <math>p_0,\!</math> scanner <math>\mathrm{H}</math> is reading cell <math>r_1,\!</math> and</p>
  At the time p_0, cell r_2 contains "#", and
+
<p>At the time <math>p_0,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
<p>At the time <math>p_0,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{1},</math> and</p>
 +
<p>At the time <math>p_0,\!</math> cell <math>r_2\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
|-
 +
|
 +
<p>At the time <math>p_1,\!</math> machine <math>\mathrm{M}</math> is in the state <math>q_1,\!</math> and</p>
 +
<p>At the time <math>p_1,\!</math> scanner <math>\mathrm{H}</math> is reading cell <math>r_2,\!</math> and</p>
 +
<p>At the time <math>p_1,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
<p>At the time <math>p_1,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{1},</math> and</p>
 +
<p>At the time <math>p_1,\!</math> cell <math>r_2\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
|-
 +
|
 +
<p>At the time <math>p_2,\!</math> machine <math>\mathrm{M}</math> is in the state <math>q_*,\!</math> and</p>
 +
<p>At the time <math>p_2,\!</math> scanner <math>\mathrm{H}</math> is reading cell <math>r_1,\!</math> and</p>
 +
<p>At the time <math>p_2,\!</math> cell <math>r_0\!</math> contains the symbol <math>\texttt{\#},</math> and</p>
 +
<p>At the time <math>p_2,\!</math> cell <math>r_1\!</math> contains the symbol <math>\texttt{1},</math> and</p>
 +
<p>At the time <math>p_2,\!</math> cell <math>r_2\!</math> contains the symbol <math>\texttt{\#}.</math></p>
 +
|}
  
  At the time p_1, M is in the state q_1, and
+
The output of <math>\mathrm{Stunt}(2)</math> being the symbol that rests under the tape head <math>\mathrm{H}</math> when and if the machine <math>\mathrm{M}</math> reaches one of its resting states, we get the result that <math>\mathrm{Parity}(1) = 1.</math>
  At the time p_1, H is reading cell r_2, and
 
  At the time p_1, cell r_0 contains "#", and
 
  At the time p_1, cell r_1 contains "1", and
 
  At the time p_1, cell r_2 contains "#", and
 
  
  At the time p_2, M is in the state q_*, and
+
==Document History==
  At the time p_2, H is reading cell r_1, and
 
  At the time p_2, cell r_0 contains "#", and
 
  At the time p_2, cell r_1 contains "1", and
 
  At the time p_2, cell r_2 contains "#".
 
 
 
The output of Stunt(2) being the symbol that rests under
 
the tape head H when and if the machine M reaches one of
 
its resting states, we get the result that Parity(1) = 1.
 
</pre>
 
 
 
==Work Area==
 
  
<pre>
+
===Ontology List : Feb&ndash;Mar 2004===
DATA 20.  http://forum.wolframscience.com/showthread.php?postid=791#post791
 
  
Let's see how this information about the transformation F,
+
* http://suo.ieee.org/ontology/thrd2.html#05457
arrived at by eyeballing the raw data, comports with what
+
# http://suo.ieee.org/ontology/msg05457.html
we derived through a more systematic symbolic computation.
+
# http://suo.ieee.org/ontology/msg05458.html
 +
# http://suo.ieee.org/ontology/msg05459.html
 +
# http://suo.ieee.org/ontology/msg05460.html
 +
# http://suo.ieee.org/ontology/msg05461.html
 +
# http://suo.ieee.org/ontology/msg05462.html
 +
# http://suo.ieee.org/ontology/msg05463.html
 +
# http://suo.ieee.org/ontology/msg05464.html
 +
# http://suo.ieee.org/ontology/msg05465.html
 +
# http://suo.ieee.org/ontology/msg05466.html
 +
# http://suo.ieee.org/ontology/msg05467.html
 +
# http://suo.ieee.org/ontology/msg05469.html
 +
# http://suo.ieee.org/ontology/msg05470.html
 +
# http://suo.ieee.org/ontology/msg05471.html
 +
# http://suo.ieee.org/ontology/msg05472.html
 +
# http://suo.ieee.org/ontology/msg05473.html
 +
# http://suo.ieee.org/ontology/msg05474.html
 +
# http://suo.ieee.org/ontology/msg05475.html
 +
# http://suo.ieee.org/ontology/msg05476.html
 +
# http://suo.ieee.org/ontology/msg05479.html
  
The results of the various operator actions that we have just
+
===NKS Forum : Feb&ndash;Jun 2004===
computed are summarized in Tables 66-i and 66-ii from my paper,
 
and I have attached these as a text file below.
 
  
Table 66-i. Computation Summary for f<u, v> = ((u)(v))
+
* http://forum.wolframscience.com/archive/topic/228-1.html
o--------------------------------------------------------------------------------o
+
* http://forum.wolframscience.com/showthread.php?threadid=228
|                                                                                |
+
* http://forum.wolframscience.com/printthread.php?threadid=228&perpage=50
| !e!f  = uv.   1      + u(v).   1      + (u)v.   1      + (u)(v).   0      |
+
# http://forum.wolframscience.com/showthread.php?postid=664#post664
|                                                                                |
+
# http://forum.wolframscience.com/showthread.php?postid=666#post666
|  Ef  = uv. (du  dv)  + u(v). (du (dv)) + (u)v.((du) dv)  + (u)(v).((du)(dv)) |
+
# http://forum.wolframscience.com/showthread.php?postid=677#post677
|                                                                                |
+
# http://forum.wolframscience.com/showthread.php?postid=684#post684
|  Df  = uv. du  dv  + u(v). du (dv)  + (u)v. (du) dv  + (u)(v).((du)(dv)) |
+
# http://forum.wolframscience.com/showthread.php?postid=689#post689
|                                                                                |
+
# http://forum.wolframscience.com/showthread.php?postid=697#post697
|  df  = uv.   0      + u(v). du      + (u)v.     dv  + (u)(v). (du, dv)  |
+
# http://forum.wolframscience.com/showthread.php?postid=708#post708
|                                                                                |
+
# http://forum.wolframscience.com/showthread.php?postid=721#post721
|  rf  = uv. du  dv  + u(v). du  dv  + (u)v. du  dv  + (u)(v). du  dv  |
+
# http://forum.wolframscience.com/showthread.php?postid=722#post722
|                                                                                |
+
# http://forum.wolframscience.com/showthread.php?postid=725#post725
o--------------------------------------------------------------------------------o
+
# http://forum.wolframscience.com/showthread.php?postid=733#post733
 +
# http://forum.wolframscience.com/showthread.php?postid=756#post756
 +
# http://forum.wolframscience.com/showthread.php?postid=759#post759
 +
# http://forum.wolframscience.com/showthread.php?postid=764#post764
 +
# http://forum.wolframscience.com/showthread.php?postid=766#post766
 +
# http://forum.wolframscience.com/showthread.php?postid=767#post767
 +
# http://forum.wolframscience.com/showthread.php?postid=773#post773
 +
# http://forum.wolframscience.com/showthread.php?postid=775#post775
 +
# http://forum.wolframscience.com/showthread.php?postid=777#post777
 +
# http://forum.wolframscience.com/showthread.php?postid=791#post791
 +
# http://forum.wolframscience.com/showthread.php?postid=1458#post1458
 +
# http://forum.wolframscience.com/showthread.php?postid=1461#post1461
 +
# http://forum.wolframscience.com/showthread.php?postid=1463#post1463
 +
# http://forum.wolframscience.com/showthread.php?postid=1464#post1464
 +
# http://forum.wolframscience.com/showthread.php?postid=1467#post1467
 +
# http://forum.wolframscience.com/showthread.php?postid=1469#post1469
 +
# http://forum.wolframscience.com/showthread.php?postid=1470#post1470
 +
# http://forum.wolframscience.com/showthread.php?postid=1471#post1471
 +
# http://forum.wolframscience.com/showthread.php?postid=1473#post1473
 +
# http://forum.wolframscience.com/showthread.php?postid=1475#post1475
 +
# http://forum.wolframscience.com/showthread.php?postid=1479#post1479
 +
# http://forum.wolframscience.com/showthread.php?postid=1489#post1489
 +
# http://forum.wolframscience.com/showthread.php?postid=1490#post1490
  
Table 66-ii.  Computation Summary for g<u, v> = ((u, v))
+
===Inquiry List : Feb&ndash;Jun 2004===
o--------------------------------------------------------------------------------o
 
|                                                                                |
 
| !e!g  = uv.    1      + u(v).    0      + (u)v.    0      + (u)(v).    1      |
 
|                                                                                |
 
|  Eg  = uv.((du, dv)) + u(v). (du, dv)  + (u)v. (du, dv)  + (u)(v).((du, dv)) |
 
|                                                                                |
 
|  Dg  = uv. (du, dv)  + u(v). (du, dv)  + (u)v. (du, dv)  + (u)(v). (du, dv)  |
 
|                                                                                |
 
|  dg  = uv. (du, dv)  + u(v). (du, dv)  + (u)v. (du, dv)  + (u)(v). (du, dv)  |
 
|                                                                                |
 
|  rg  = uv.    0      + u(v).    0      + (u)v.    0      + (u)(v).    0      |
 
|                                                                                |
 
o--------------------------------------------------------------------------------o
 
  
 +
* http://stderr.org/pipermail/inquiry/2004-February/thread.html#1228
 +
* http://stderr.org/pipermail/inquiry/2004-March/thread.html#1235
 +
* http://stderr.org/pipermail/inquiry/2004-March/thread.html#1240
 +
* http://stderr.org/pipermail/inquiry/2004-June/thread.html#1630
 +
# http://stderr.org/pipermail/inquiry/2004-February/001228.html
 +
# http://stderr.org/pipermail/inquiry/2004-February/001230.html
 +
# http://stderr.org/pipermail/inquiry/2004-February/001231.html
 +
# http://stderr.org/pipermail/inquiry/2004-February/001232.html
 +
# http://stderr.org/pipermail/inquiry/2004-February/001233.html
 +
# http://stderr.org/pipermail/inquiry/2004-February/001234.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001235.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001236.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001237.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001238.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001240.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001242.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001243.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001244.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001245.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001246.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001247.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001248.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001249.html
 +
# http://stderr.org/pipermail/inquiry/2004-March/001255.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001630.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001631.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001632.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001633.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001634.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001635.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001636.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001637.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001638.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001639.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001640.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001641.html
 +
# http://stderr.org/pipermail/inquiry/2004-June/001642.html
  
o---------------------------------------o
+
[[Category:Automata Theory]]
|                                      |
+
[[Category:Computation]]
|                  o                  |
+
[[Category:Computational Complexity]]
|                  / \                  |
+
[[Category:Computer Science]]
|                /  \                |
+
[[Category:Differential Logic]]
|                /    \                |
+
[[Category:Formal Languages]]
|              o      o              |
+
[[Category:Graph Theory]]
|              / \    / \              |
+
[[Category:Logic]]
|            /  \  /  \            |
+
[[Category:Logical Graphs]]
|            /    \ /    \            |
+
[[Category:Mathematics]]
|          o      o      o          |
+
[[Category:Programming Languages]]
|          / \    / \    / \          |
+
[[Category:Turing Machines]]
|        /  \  /  \  /  \        |
+
[[Category:Visualization]]
|        /    \ /    \ /    \        |
 
|      o      o      o      o      |
 
|      / \    / \    / \    / \      |
 
|    /  \  /  \  /  \  /  \    |
 
|    /    \ /    \ /    \ /    \    |
 
|  o      o      o      o      o  |
 
|  |\    / \    / \    / \    /|  |
 
|  | \  /  \  /  \  /  \  / |  |
 
|  |  \ /    \ /    \ /    \ /  |  |
 
|  |  o      o      o      o  |  |
 
|  |  |\    / \    / \    /|  |  |
 
|  |  | \  /  \  /  \  / |  |  |
 
|  | u |  \ /    \ /    \ /  | v |  |
 
|  o---+---o      o      o---+---o  |
 
|      |    \    / \    /    |      |
 
|      |    \  /  \  /    |      |
 
|      | du  \ /    \ /  dv |      |
 
|      o-------o      o-------o      |
 
|                \    /                |
 
|                \  /                |
 
|                  \ /                  |
 
|                  o                  |
 
|                                      |
 
o---------------------------------------o
 
</pre>
 
 
 
==Discussion==
 
 
 
<pre>
 
PD = Philip Dutton
 
 
 
PD: I've been watching your posts.
 
 
 
PD: I am not an expert on logic infrastructures but I find the posts
 
    interesting (despite not understanding much of it).  I am like
 
    the diagrams.  I have recently been trying to understand CA's
 
    using a particular perspective: sinks and sources.  I think
 
    that all CA's are simply combinations of sinks and sources.
 
    How they interact (or intrude into each other's domains)
 
    would most likely be a result of the rules (and initial
 
    configuration of on or off cells).
 
 
 
PD: Anyway, to be short, I "see" diamond shapes quite often in
 
    your diagrams.  Triangles (either up or down) or diamonds
 
    (combination of an up and down triangle) make me think
 
    soley of sinks and sources.  I think of the diamond to
 
    be a source which, during the course of progression,
 
    is expanding (because it is producing) and then starts
 
    to act as a sink  (because it consumes) -- and hence the
 
    diamond.  I can't stop thinking about sinks and sources in
 
    CA's and so I thought I would ask you if there is some way
 
    to tie the two worlds together (CA's of sinks and sources
 
    together with your differential constructs).
 
 
 
PD: Any thoughts?
 
 
 
Yes, I'm hoping that there's a lot of stuff analogous to
 
R-world dynamics to be discovered in this B-world variety,
 
indeed, that's kind of why I set out on this investigation --
 
oh, gee, has it been that long? -- I guess about 1989 or so,
 
when I started to see this "differential logic" angle on what
 
I had previously studied in systems theory as the "qualitative
 
approach to differential equations".  I think we used to use the
 
words "attractor" and "basin" more often than "sink", but a source
 
is still a source as time goes by, and I do remember using the word
 
"sink" a lot when I was a freshperson in physics, before I got logic.
 
 
 
I have spent the last 15 years doing a funny mix of practice in stats
 
and theory in math, but I did read early works by Von Neumann, Burks,
 
Ulam, and later stuff by Holland on CA's.  Still, it may be a while
 
before I have re-heated my concrete intuitions about them in the
 
NKS way of thinking.
 
 
 
There are some fractal-looking pictures that emerge when
 
I turn to "higher order propositional expressions" (HOPE's).
 
I have discussed this topic elswhere on the web and can look
 
it up now if your are interested, but I am trying to make my
 
e-positions somewhat clearer for the NKS forum than I have
 
tried to do before.
 
 
 
But do not hestitate to dialogue all this stuff on the boards,
 
as that's what always seems to work the best.  What I've found
 
works best for me, as I can hardly remember what I was writing
 
last month without Google, is to archive a copy at one of the
 
other Google-visible discussion lists that I'm on at present.
 
</pre>
 
 
 
==Document History==
 
 
 
<pre>
 
DATA.  Differential Analytic Turing Automata
 
 
 
Ontology List
 
 
 
01.  http://suo.ieee.org/ontology/msg05457.html
 
02.  http://suo.ieee.org/ontology/msg05458.html
 
03.  http://suo.ieee.org/ontology/msg05459.html
 
04.  http://suo.ieee.org/ontology/msg05460.html
 
05.  http://suo.ieee.org/ontology/msg05461.html
 
06.  http://suo.ieee.org/ontology/msg05462.html
 
07.  http://suo.ieee.org/ontology/msg05463.html
 
08.  http://suo.ieee.org/ontology/msg05464.html
 
09.  http://suo.ieee.org/ontology/msg05465.html
 
10.  http://suo.ieee.org/ontology/msg05466.html
 
11.  http://suo.ieee.org/ontology/msg05467.html
 
12.  http://suo.ieee.org/ontology/msg05469.html
 
13.  http://suo.ieee.org/ontology/msg05470.html
 
14.  http://suo.ieee.org/ontology/msg05471.html
 
15.  http://suo.ieee.org/ontology/msg05472.html
 
16.  http://suo.ieee.org/ontology/msg05473.html
 
17.  http://suo.ieee.org/ontology/msg05474.html
 
18.  http://suo.ieee.org/ontology/msg05475.html
 
19.  http://suo.ieee.org/ontology/msg05476.html
 
20.  http://suo.ieee.org/ontology/msg05479.html
 
 
 
Inquiry List
 
 
 
00.  http://stderr.org/pipermail/inquiry/2004-February/thread.html#1228
 
00.  http://stderr.org/pipermail/inquiry/2004-March/thread.html#1235
 
00.  http://stderr.org/pipermail/inquiry/2004-March/thread.html#1240
 
00.  http://stderr.org/pipermail/inquiry/2004-June/thread.html#1630
 
 
 
01.  http://stderr.org/pipermail/inquiry/2004-February/001228.html
 
02.  http://stderr.org/pipermail/inquiry/2004-February/001230.html
 
03.  http://stderr.org/pipermail/inquiry/2004-February/001231.html
 
04.  http://stderr.org/pipermail/inquiry/2004-February/001232.html
 
05.  http://stderr.org/pipermail/inquiry/2004-February/001233.html
 
06.  http://stderr.org/pipermail/inquiry/2004-February/001234.html
 
07.  http://stderr.org/pipermail/inquiry/2004-March/001235.html
 
08.  http://stderr.org/pipermail/inquiry/2004-March/001236.html
 
09.  http://stderr.org/pipermail/inquiry/2004-March/001237.html
 
10.  http://stderr.org/pipermail/inquiry/2004-March/001238.html
 
11.  http://stderr.org/pipermail/inquiry/2004-March/001240.html
 
12.  http://stderr.org/pipermail/inquiry/2004-March/001242.html
 
13.  http://stderr.org/pipermail/inquiry/2004-March/001243.html
 
14.  http://stderr.org/pipermail/inquiry/2004-March/001244.html
 
15.  http://stderr.org/pipermail/inquiry/2004-March/001245.html
 
16.  http://stderr.org/pipermail/inquiry/2004-March/001246.html
 
17.  http://stderr.org/pipermail/inquiry/2004-March/001247.html
 
18.  http://stderr.org/pipermail/inquiry/2004-March/001248.html
 
19.  http://stderr.org/pipermail/inquiry/2004-March/001249.html
 
20.  http://stderr.org/pipermail/inquiry/2004-March/001255.html
 
21.  http://stderr.org/pipermail/inquiry/2004-June/001630.html
 
22.  http://stderr.org/pipermail/inquiry/2004-June/001631.html
 
23.  http://stderr.org/pipermail/inquiry/2004-June/001632.html
 
24.  http://stderr.org/pipermail/inquiry/2004-June/001633.html
 
25.  http://stderr.org/pipermail/inquiry/2004-June/001634.html
 
26.  http://stderr.org/pipermail/inquiry/2004-June/001635.html
 
27.  http://stderr.org/pipermail/inquiry/2004-June/001636.html
 
28.  http://stderr.org/pipermail/inquiry/2004-June/001637.html
 
29.  http://stderr.org/pipermail/inquiry/2004-June/001638.html
 
30.  http://stderr.org/pipermail/inquiry/2004-June/001639.html
 
31.  http://stderr.org/pipermail/inquiry/2004-June/001640.html
 
32.  http://stderr.org/pipermail/inquiry/2004-June/001641.html
 
33.  http://stderr.org/pipermail/inquiry/2004-June/001642.html
 
 
 
NKS Forum
 
 
 
00.  http://forum.wolframscience.com/showthread.php?threadid=228
 
01.  http://forum.wolframscience.com/showthread.php?postid=664#post664
 
02.  http://forum.wolframscience.com/showthread.php?postid=666#post666
 
03.  http://forum.wolframscience.com/showthread.php?postid=677#post677
 
04.  http://forum.wolframscience.com/showthread.php?postid=684#post684
 
05.  http://forum.wolframscience.com/showthread.php?postid=689#post689
 
06.  http://forum.wolframscience.com/showthread.php?postid=697#post697
 
07.  http://forum.wolframscience.com/showthread.php?postid=708#post708
 
08.  http://forum.wolframscience.com/showthread.php?postid=721#post721
 
09.  http://forum.wolframscience.com/showthread.php?postid=722#post722
 
10.  http://forum.wolframscience.com/showthread.php?postid=725#post725
 
11.  http://forum.wolframscience.com/showthread.php?postid=733#post733
 
12.  http://forum.wolframscience.com/showthread.php?postid=756#post756
 
13.  http://forum.wolframscience.com/showthread.php?postid=759#post759
 
14.  http://forum.wolframscience.com/showthread.php?postid=764#post764
 
15.  http://forum.wolframscience.com/showthread.php?postid=766#post766
 
16.  http://forum.wolframscience.com/showthread.php?postid=767#post767
 
17.  http://forum.wolframscience.com/showthread.php?postid=773#post773
 
18.  http://forum.wolframscience.com/showthread.php?postid=775#post775
 
19.  http://forum.wolframscience.com/showthread.php?postid=777#post777
 
20.  http://forum.wolframscience.com/showthread.php?postid=791#post791
 
21.  http://forum.wolframscience.com/showthread.php?postid=1458#post1458
 
22.  http://forum.wolframscience.com/showthread.php?postid=1461#post1461
 
23.  http://forum.wolframscience.com/showthread.php?postid=1463#post1463
 
24.  http://forum.wolframscience.com/showthread.php?postid=1464#post1464
 
25.  http://forum.wolframscience.com/showthread.php?postid=1467#post1467
 
26.  http://forum.wolframscience.com/showthread.php?postid=1469#post1469
 
27.  http://forum.wolframscience.com/showthread.php?postid=1470#post1470
 
28.  http://forum.wolframscience.com/showthread.php?postid=1471#post1471
 
29.  http://forum.wolframscience.com/showthread.php?postid=1473#post1473
 
30.  http://forum.wolframscience.com/showthread.php?postid=1475#post1475
 
31.  http://forum.wolframscience.com/showthread.php?postid=1479#post1479
 
32.  http://forum.wolframscience.com/showthread.php?postid=1489#post1489
 
33.  http://forum.wolframscience.com/showthread.php?postid=1490#post1490
 
</pre>
 

Latest revision as of 15:30, 11 October 2013

Author: Jon Awbrey

The task ahead is to chart a course from general ideas about transformational equivalence classes of graphs to a notion of differential analytic turing automata (DATA). It may be a while before we get within sight of that goal, but it will provide a better measure of motivation to name the thread after the envisioned end rather than the more homely starting place.

The basic idea is as follows. One has a set \(\mathcal{G}\) of graphs and a set \(\mathcal{T}\) of transformation rules, and each rule \(\mathrm{t} \in \mathcal{T}\) has the effect of transforming graphs into graphs, \(\mathrm{t} : \mathcal{G} \to \mathcal{G}.\) In the cases that we shall be studying, this set of transformation rules partitions the set of graphs into transformational equivalence classes (TECs).

There are many interesting excursions to be had here, but I will focus mainly on logical applications, and and so the TECs I talk about will almost always have the character of logical equivalence classes (LECs).

An example that will figure heavily in the sequel is given by rooted trees as the species of graphs and a pair of equational transformation rules that derive from the graphical calculi of C.S. Peirce, as revived and extended by George Spencer Brown.

Here are the fundamental transformation rules, also referred to as the arithmetic axioms, more precisely, the arithmetic initials.

PERS Figure 01.jpg (1)
PERS Figure 02.jpg (2)

That should be enough to get started.

Cactus Language

I will be making use of the cactus language extension of Peirce's Alpha Graphs, so called because it uses a species of graphs that are usually called "cacti" in graph theory. The last exposition of the cactus syntax that I've written can be found here:

The representational and computational efficiency of the cactus language for the tasks that are usually associated with boolean algebra and propositional calculus makes it possible to entertain a further extension, to what we may call differential logic, because it develops this basic level of logic in the same way that differential calculus augments analytic geometry to handle change and diversity. There are several different introductions to differential logic that I have written and distributed across the Internet. You might start with the following couple of treatments:

I will draw on those previously advertised resources of notation and theory as needed, but right now I sense the need for some concrete examples.

Example 1

Let's say we have a system that is known by the name of its state space \(X\!\) and we have a boolean state variable \(x : X \to \mathbb{B},\!\) where \(\mathbb{B} = \{ 0, 1 \}.\!\)

We observe \(X\!\) for a while, relative to a discrete time frame, and we write down the following sequence of values for \(x.\!\)

\(\begin{array}{c|c} t & x \'"`UNIQ-MathJax1-QINU`"' in other words'"`UNIQ-MathJax2-QINU`"' =='"`UNIQ--h-10--QINU`"'Notions of Approximation== {| cellpadding="2" cellspacing="2" width="100%" | width="60%" |   | width="40%" | for equalities are so weighed<br> that curiosity in neither can<br> make choice of either's moiety. |- | height="50px" |   | valign="top" | — ''King Lear'', Sc.1.5–7 (Quarto) |- |   | for qualities are so weighed<br> that curiosity in neither can<br> make choice of either's moiety.<br> |- | height="50px" |   | valign="top" | — ''King Lear'', 1.1.5–6 (Folio) |} Justifying a notion of approximation is a little more involved in general, and especially in these discrete logical spaces, than it would be expedient for people in a hurry to tangle with right now. I will just say that there are ''naive'' or ''obvious'' notions and there are ''sophisticated'' or ''subtle'' notions that we might choose among. The later would engage us in trying to construct proper logical analogues of Lie derivatives, and so let's save that for when we have become subtle or sophisticated or both. Against or toward that day, as you wish, let's begin with an option in plain view. Figure 1.4 illustrates one way of ranging over the cells of the underlying universe \(U^\bullet = [u, v]\!\) and selecting at each cell the linear proposition in \(\mathrm{d}U^\bullet = [\mathrm{d}u, \mathrm{d}v]\!\) that best approximates the patch of the difference map \({\mathrm{D}f}\!\) that is located there, yielding the following formula for the differential \(\mathrm{d}f.\!\)

\(\begin{array}{*{11}{c}} \mathrm{d}f & = & \mathrm{d}\texttt{((} u \texttt{)(} v \texttt{))} & = & uv \cdot 0 & + & u \texttt{(} v \texttt{)} \cdot \mathrm{d}u & + & \texttt{(} u \texttt{)} v \cdot \mathrm{d}v & + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)} \end{array}\)

o---------------------------------------o
|                                       |
|                   o                   |
|                  / \                  |
|                 /   \                 |
|                /     \                |
|               o       o               |
|              / \     / \              |
|             /   \   /   \             |
|            /     \ /     \            |
|           o       o       o           |
|          / \     / \     / \          |
|         /   \   /   \   /   \         |
|        /     \ /     \ /     \        |
|       o       o       o       o       |
|      / \     /%\     /%\     / \      |
|     /   \   /%%%\   /%%%\   /   \     |
|    /     \ /%%%%%\ /%%%%%\ /     \    |
|   o       o%%%%%%%o%%%%%%%o       o   |
|   |\     /%\%%%%%/ \%%%%%/%\     /|   |
|   | \   /%%%\%%%/   \%%%/%%%\   / |   |
|   |  \ /%%%%%\%/     \%/%%%%%\ /  |   |
|   |   o%%%%%%%o       o%%%%%%%o   |   |
|   |   |\%%%%%/%\     /%\%%%%%/|   |   |
|   |   | \%%%/%%%\   /%%%\%%%/ |   |   |
|   | u |  \%/%%%%%\ /%%%%%\%/  | v |   |
|   o---+---o%%%%%%%o%%%%%%%o---+---o   |
|       |    \%%%%%/ \%%%%%/    |       |
|       |     \%%%/   \%%%/     |       |
|       | du   \%/     \%/   dv |       |
|       o-------o       o-------o       |
|                \     /                |
|                 \   /                 |
|                  \ /                  |
|                   o                   |
|                                       |
o---------------------------------------o
Figure 1.4.  df = linear approx to Df

Figure 2.4 illustrates one way of ranging over the cells of the underlying universe \(U^\bullet = [u, v]\!\) and selecting at each cell the linear proposition in \(\mathrm{d}U^\bullet = [du, dv]\!\) that best approximates the patch of the difference map \(\mathrm{D}g\!\) that is located there, yielding the following formula for the differential \(\mathrm{d}g.\!\)

\(\begin{array}{*{11}{c}} \mathrm{d}g & = & \mathrm{d}\texttt{((} u \texttt{,} v \texttt{))} & = & uv \cdot \texttt{(} \mathrm{d}u \texttt{,} \mathrm{d}v \texttt{)} & + & u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,} \mathrm{d}v \texttt{)} & + & \texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{,} \mathrm{d}v \texttt{)} & + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,} \mathrm{d}v \texttt{)} \end{array}\)

o---------------------------------------o
|                                       |
|                   o                   |
|                  / \                  |
|                 /   \                 |
|                /     \                |
|               o       o               |
|              /%\     /%\              |
|             /%%%\   /%%%\             |
|            /%%%%%\ /%%%%%\            |
|           o%%%%%%%o%%%%%%%o           |
|          /%\%%%%%/ \%%%%%/%\          |
|         /%%%\%%%/   \%%%/%%%\         |
|        /%%%%%\%/     \%/%%%%%\        |
|       o%%%%%%%o       o%%%%%%%o       |
|      / \%%%%%/ \     / \%%%%%/ \      |
|     /   \%%%/   \   /   \%%%/   \     |
|    /     \%/     \ /     \%/     \    |
|   o       o       o       o       o   |
|   |\     /%\     / \     /%\     /|   |
|   | \   /%%%\   /   \   /%%%\   / |   |
|   |  \ /%%%%%\ /     \ /%%%%%\ /  |   |
|   |   o%%%%%%%o       o%%%%%%%o   |   |
|   |   |\%%%%%/%\     /%\%%%%%/|   |   |
|   |   | \%%%/%%%\   /%%%\%%%/ |   |   |
|   | u |  \%/%%%%%\ /%%%%%\%/  | v |   |
|   o---+---o%%%%%%%o%%%%%%%o---+---o   |
|       |    \%%%%%/ \%%%%%/    |       |
|       |     \%%%/   \%%%/     |       |
|       | du   \%/     \%/   dv |       |
|       o-------o       o-------o       |
|                \     /                |
|                 \   /                 |
|                  \ /                  |
|                   o                   |
|                                       |
o---------------------------------------o
Figure 2.4.  dg = linear approx to Dg

Well, \(g,\!\) that was easy, seeing as how \(\mathrm{D}g\!\) is already linear at each locus, \(\mathrm{d}g = \mathrm{D}g.\!\)

Analytic Series

We have been conducting the differential analysis of the logical transformation \(F : [u, v] \mapsto [u, v]\!\) defined as \(F : (u, v) \mapsto ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~ ),\!\) and this means starting with the extended transformation \(\mathrm{E}F : [u, v, \mathrm{d}u, \mathrm{d}v] \to [u, v, \mathrm{d}u, \mathrm{d}v]\!\) and breaking it into an analytic series, \(\mathrm{E}F = F + \mathrm{d}F + \mathrm{d}^2 F + \ldots,\!\) and so on until there is nothing left to analyze any further.

As a general rule, one proceeds by way of the following stages:

\(\begin{array}{*{6}{l}} 1. & \mathrm{E}F & = & \mathrm{d}^0 F & + & \mathrm{r}^0 F \\ 2. & \mathrm{r}^0 F & = & \mathrm{d}^1 F & + & \mathrm{r}^1 F \\ 3. & \mathrm{r}^1 F & = & \mathrm{d}^2 F & + & \mathrm{r}^2 F \\ 4. & \ldots \end{array}\)

In our analysis of the transformation \(F,\!\) we carried out Step 1 in the more familiar form \(\mathrm{E}F = F + \mathrm{D}F\!\) and we have just reached Step 2 in the form \(\mathrm{D}F = \mathrm{d}F + \mathrm{r}F,\!\) where \(\mathrm{r}F\!\) is the residual term that remains for us to examine next.

Note. I'm am trying to give quick overview here, and this forces me to omit many picky details. The picky reader may wish to consult the more detailed presentation of this material at the following locations:

Let's push on with the analysis of the transformation:

\(\begin{matrix} F & : & (u, v) & \mapsto & (f(u, v), g(u, v)) & = & ( ~ \texttt{((} u \texttt{)(} v \texttt{))} ~,~ \texttt{((} u \texttt{,~} v \texttt{))} ~) \end{matrix}\)

For ease of comparison and computation, I will collect the Figures that we need for the remainder of the work together on one page.

Computation Summary for Logical Disjunction

Figure 1.1 shows the expansion of \(f = \texttt{((} u \texttt{)(} v \texttt{))}\!\) over \([u, v]\!\) to produce the expression:

\(\begin{matrix} uv & + & u \texttt{(} v \texttt{)} & + & \texttt{(} u \texttt{)} v \end{matrix}\)

Figure 1.2 shows the expansion of \(\mathrm{E}f = \texttt{((} u + \mathrm{d}u \texttt{)(} v + \mathrm{d}v \texttt{))}\!\) over \([u, v]\!\) to produce the expression:

\(\begin{matrix} uv \cdot \texttt{(} \mathrm{d}u ~ \mathrm{d}v \texttt{)} & + & u \texttt{(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{(} \mathrm{d}v \texttt{))} & + & \texttt{(} u \texttt{)} v \cdot \texttt{((} \mathrm{d}u \texttt{)} \mathrm{d}v \texttt{)} & + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{((} \mathrm{d}u \texttt{)(} \mathrm{d}v \texttt{))} \end{matrix}\)

In general, \(\mathrm{E}f\!\) tells you what you would have to do, from wherever you are in the universe \([u, v],\!\) if you want to end up in a place where \(f\!\) is true. In this case, where the prevailing proposition \(f\!\) is \(\texttt{((} u \texttt{)(} v \texttt{))},\!\) the indication \(uv \cdot \texttt{(} \mathrm{d}u ~ \mathrm{d}v \texttt{)}\!\) of \(\mathrm{E}f\!\) tells you this: If \(u\!\) and \(v\!\) are both true where you are, then just don't change both \(u\!\) and \(v\!\) at once, and you will end up in a place where \(\texttt{((} u \texttt{)(} v \texttt{))}\!\) is true.

Figure 1.3 shows the expansion of \(\mathrm{D}f\) over \([u, v]\!\) to produce the expression:

\(\begin{matrix} uv \cdot \mathrm{d}u ~ \mathrm{d}v & + & u \texttt{(} v \texttt{)} \cdot \mathrm{d}u \texttt{(} \mathrm{d}v \texttt{)} & + & \texttt{(} u \texttt{)} v \cdot \texttt{(} \mathrm{d}u \texttt{)} \mathrm{d}v & + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{((} \mathrm{d}u \texttt{)(} \mathrm{d}v \texttt{))} \end{matrix}\)

In general, \({\mathrm{D}f}\!\) tells you what you would have to do, from wherever you are in the universe \([u, v],\!\) if you want to bring about a change in the value of \(f,\!\) that is, if you want to get to a place where the value of \(f\!\) is different from what it is where you are. In the present case, where the reigning proposition \(f\!\) is \(\texttt{((} u \texttt{)(} v \texttt{))},\!\) the term \(uv \cdot \mathrm{d}u ~ \mathrm{d}v\!\) of \({\mathrm{D}f}\!\) tells you this: If \(u\!\) and \(v\!\) are both true where you are, then you would have to change both \(u\!\) and \(v\!\) in order to reach a place where the value of \(f\!\) is different from what it is where you are.

Figure 1.4 approximates \({\mathrm{D}f}\!\) by the linear form \(\mathrm{d}f\!\) that expands over \([u, v]\!\) as follows:

\(\begin{matrix} \mathrm{d}f & = & uv \cdot 0 & + & u \texttt{(} v \texttt{)} \cdot \mathrm{d}u & + & \texttt{(} u \texttt{)} v \cdot \mathrm{d}v & + & \texttt{(} u \texttt{)(} v \texttt{)} \cdot \texttt{(} \mathrm{d}u \texttt{,~} \mathrm{d}v \texttt{)} \'"`UNIQ-MathJax3-QINU`"' Another way to convey the same information is by means of the extended proposition'"`UNIQ-MathJax4-QINU`"' {| align="center" cellpadding="8" style="text-align:center" | \(\text{Orbit 2}\!\)

\(\begin{array}{c|cc|cc|cc|} t & u & v & \mathrm{d}u & \mathrm{d}v & \mathrm{d}^2 u & \mathrm{d}^2 v \\[8pt] 0 & 0 & 0 & 0 & 1 & 1 & 0 \\ 1 & 0 & 1 & 1 & 1 & 1 & 1 \\ 2 & 1 & 0 & 0 & 0 & 0 & 0 \\ 3 & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel & {}^\shortparallel \end{array}\)

A more fine combing of the second Table brings to mind a rule that partly covers the remaining cases, that is, \(\mathrm{d}u = v, ~ \mathrm{d}v = \texttt{(} u \texttt{)}.\!\) This much information about Orbit 2 is also encapsulated by the extended proposition \(\texttt{(} uv \texttt{)((} \mathrm{d}u \texttt{,} v \texttt{))(} \mathrm{d}v, u \texttt{)},\!\) which says that \(u\!\) and \(v\!\) are not both true at the same time, while \(\mathrm{d}u\!\) is equal in value to \(v\!\) and \(\mathrm{d}v\!\) is opposite in value to \(u.\!\)

Turing Machine Example

See Theme One Program for documentation of the cactus graph syntax and the propositional modeling program used below.

By way of providing a simple illustration of Cook's Theorem, namely, that “Propositional Satisfiability is NP-Complete”, I will describe one way to translate finite approximations of turing machines into propositional expressions, using the cactus language syntax for propositional calculus that I will describe in more detail as we proceed.

\(\mathrm{Stilt}(k) =\!\)
Space and time limited turing machine,
with \(k\!\) units of space and \(k\!\) units of time.
\(\mathrm{Stunt}(k) =\!\)
Space and time limited turing machine,
for computing the parity of a bit string, with number of tape cells of input equal to \(k.\!\)

I will follow the pattern of discussion in Herbert Wilf (1986), Algorithms and Complexity, pp. 188–201, but translate his logical formalism into cactus language, which is more efficient in regard to the number of propositional clauses that are required.

A turing machine for computing the parity of a bit string is described by means of the following Figure and Table.

Parity Machine.jpg
\(\text{Figure 3.} ~~ \text{Parity Machine}\!\)


Table 4.  Parity Machine
o-------o--------o-------------o---------o------------o
| State | Symbol | Next Symbol | Ratchet | Next State |
|   Q   |   S    |     S'      |   dR    |     Q'     |
o-------o--------o-------------o---------o------------o
|   0   |   0    |     0       |   +1    |     0      |
|   0   |   1    |     1       |   +1    |     1      |
|   0   |   #    |     #       |   -1    |     #      |
|   1   |   0    |     0       |   +1    |     1      |
|   1   |   1    |     1       |   +1    |     0      |
|   1   |   #    |     #       |   -1    |     *      |
o-------o--------o-------------o---------o------------o


The TM has a finite automaton (FA) as one component. Let us refer to this particular FA by the name of \(\mathrm{M}.\)

The tape head (that is, the read unit) will be called \(\mathrm{H}.\) The registers are also called tape cells or tape squares.

Finite Approximations

To see how each finite approximation to a given turing machine can be given a purely propositional description, one fixes the parameter \(k\!\) and limits the rest of the discussion to describing \(\mathrm{Stilt}(k),\!\) which is not really a full-fledged TM anymore but just a finite automaton in disguise.

In this example, for the sake of a minimal illustration, we choose \(k = 2,\!\) and discuss \(\mathrm{Stunt}(2).\) Since the zeroth tape cell and the last tape cell are both occupied by the character \(^{\backprime\backprime}\texttt{\#}^{\prime\prime}\) that is used for both the beginning of file \((\mathrm{bof})\) and the end of file \((\mathrm{eof})\) markers, this allows for only one digit of significant computation.

To translate \(\mathrm{Stunt}(2)\) into propositional form we use the following collection of basic propositions, boolean variables, or logical features, depending on what one prefers to call them:

The basic propositions for describing the present state function \(\mathrm{QF} : P \to Q\) are these:

\(\begin{matrix} \texttt{p0\_q\#}, & \texttt{p0\_q*}, & \texttt{p0\_q0}, & \texttt{p0\_q1}, \\[6pt] \texttt{p1\_q\#}, & \texttt{p1\_q*}, & \texttt{p1\_q0}, & \texttt{p1\_q1}, \\[6pt] \texttt{p2\_q\#}, & \texttt{p2\_q*}, & \texttt{p2\_q0}, & \texttt{p2\_q1}, \\[6pt] \texttt{p3\_q\#}, & \texttt{p3\_q*}, & \texttt{p3\_q0}, & \texttt{p3\_q1}. \end{matrix}\)

The proposition of the form \(\texttt{pi\_qj}\) says:

At the point-in-time \(p_i,\!\) the finite state machine \(\mathrm{M}\) is in the state \(q_j.\!\)

The basic propositions for describing the present register function \(\mathrm{RF} : P \to R\) are these:

\(\begin{matrix} \texttt{p0\_r0}, & \texttt{p0\_r1}, & \texttt{p0\_r2}, & \texttt{p0\_r3}, \\[6pt] \texttt{p1\_r0}, & \texttt{p1\_r1}, & \texttt{p1\_r2}, & \texttt{p1\_r3}, \\[6pt] \texttt{p2\_r0}, & \texttt{p2\_r1}, & \texttt{p2\_r2}, & \texttt{p2\_r3}, \\[6pt] \texttt{p3\_r0}, & \texttt{p3\_r1}, & \texttt{p3\_r2}, & \texttt{p3\_r3}. \end{matrix}\)

The proposition of the form \(\texttt{pi\_rj}\) says:

At the point-in-time \(p_i,\!\) the tape head \(\mathrm{H}\) is on the tape cell \(r_j.\!\)

The basic propositions for describing the present symbol function \(\mathrm{SF} : P \to (R \to S)\) are these:

\(\begin{matrix} \texttt{p0\_r0\_s\#}, & \texttt{p0\_r0\_s*}, & \texttt{p0\_r0\_s0}, & \texttt{p0\_r0\_s1}, \\[4pt] \texttt{p0\_r1\_s\#}, & \texttt{p0\_r1\_s*}, & \texttt{p0\_r1\_s0}, & \texttt{p0\_r1\_s1}, \\[4pt] \texttt{p0\_r2\_s\#}, & \texttt{p0\_r2\_s*}, & \texttt{p0\_r2\_s0}, & \texttt{p0\_r2\_s1}, \\[4pt] \texttt{p0\_r3\_s\#}, & \texttt{p0\_r3\_s*}, & \texttt{p0\_r3\_s0}, & \texttt{p0\_r3\_s1}, \\[12pt] \texttt{p1\_r0\_s\#}, & \texttt{p1\_r0\_s*}, & \texttt{p1\_r0\_s0}, & \texttt{p1\_r0\_s1}, \\[4pt] \texttt{p1\_r1\_s\#}, & \texttt{p1\_r1\_s*}, & \texttt{p1\_r1\_s0}, & \texttt{p1\_r1\_s1}, \\[4pt] \texttt{p1\_r2\_s\#}, & \texttt{p1\_r2\_s*}, & \texttt{p1\_r2\_s0}, & \texttt{p1\_r2\_s1}, \\[4pt] \texttt{p1\_r3\_s\#}, & \texttt{p1\_r3\_s*}, & \texttt{p1\_r3\_s0}, & \texttt{p1\_r3\_s1}, \\[12pt] \texttt{p2\_r0\_s\#}, & \texttt{p2\_r0\_s*}, & \texttt{p2\_r0\_s0}, & \texttt{p2\_r0\_s1}, \\[4pt] \texttt{p2\_r1\_s\#}, & \texttt{p2\_r1\_s*}, & \texttt{p2\_r1\_s0}, & \texttt{p2\_r1\_s1}, \\[4pt] \texttt{p2\_r2\_s\#}, & \texttt{p2\_r2\_s*}, & \texttt{p2\_r2\_s0}, & \texttt{p2\_r2\_s1}, \\[4pt] \texttt{p2\_r3\_s\#}, & \texttt{p2\_r3\_s*}, & \texttt{p2\_r3\_s0}, & \texttt{p2\_r3\_s1}, \\[12pt] \texttt{p3\_r0\_s\#}, & \texttt{p3\_r0\_s*}, & \texttt{p3\_r0\_s0}, & \texttt{p3\_r0\_s1}, \\[4pt] \texttt{p3\_r1\_s\#}, & \texttt{p3\_r1\_s*}, & \texttt{p3\_r1\_s0}, & \texttt{p3\_r1\_s1}, \\[4pt] \texttt{p3\_r2\_s\#}, & \texttt{p3\_r2\_s*}, & \texttt{p3\_r2\_s0}, & \texttt{p3\_r2\_s1}, \\[4pt] \texttt{p3\_r3\_s\#}, & \texttt{p3\_r3\_s*}, & \texttt{p3\_r3\_s0}, & \texttt{p3\_r3\_s1}. \end{matrix}\)

The proposition of the form \(\texttt{pi\_rj\_sk}\) says:

At the point-in-time \(p_i,\!\) the tape cell \(r_j\!\) bears the mark \(s_k.\!\)

Initial Conditions

Given but a single free square on the tape, there are just two different sets of initial conditions for \(\mathrm{Stunt}(2),\) the finite approximation to the parity turing machine that we are presently considering.

Initial Conditions for Tape Input "0"

The following conjunction of 5 basic propositions describes the initial conditions when \(\mathrm{Stunt}(2)\) is started with an input of "0" in its free square:

\(\begin{array}{l} \texttt{p0\_q0} \\ \\ \texttt{p0\_r1} \\ \\ \texttt{p0\_r0\_s\#} \\ \texttt{p0\_r1\_s0} \\ \texttt{p0\_r2\_s\#} \end{array}\)

This conjunction of basic propositions may be read as follows:

At time \(p_0,\!\) machine \(\mathrm{M}\) is in the state \(q_0,\!\)

At time \(p_0,\!\) scanner \(\mathrm{H}\) is reading cell \(r_1,\!\)

At time \(p_0,\!\) cell \(r_0\!\) contains the symbol \(\texttt{\#},\)

At time \(p_0,\!\) cell \(r_1\!\) contains the symbol \(\texttt{0},\)

At time \(p_0,\!\) cell \(r_2\!\) contains the symbol \(\texttt{\#}.\)

Initial Conditions for Tape Input "1"

The following conjunction of 5 basic propositions describes the initial conditions when \(\mathrm{Stunt}(2)\) is started with an input of "1" in its free square:

\(\begin{array}{l} \texttt{p0\_q0} \\ \\ \texttt{p0\_r1} \\ \\ \texttt{p0\_r0\_s\#} \\ \texttt{p0\_r1\_s1} \\ \texttt{p0\_r2\_s\#} \end{array}\)

This conjunction of basic propositions may be read as follows:

At time \(p_0,\!\) machine \(\mathrm{M}\) is in the state \(q_0,\!\)

At time \(p_0,\!\) scanner \(\mathrm{H}\) is reading cell \(r_1,\!\)

At time \(p_0,\!\) cell \(r_0\!\) contains the symbol \(\texttt{\#},\)

At time \(p_0,\!\) cell \(r_1\!\) contains the symbol \(\texttt{1},\)

At time \(p_0,\!\) cell \(r_2\!\) contains the symbol \(\texttt{\#}.\)

Propositional Program

A complete description of \(\mathrm{Stunt}(2)\) in propositional form is obtained by conjoining one of the above choices for initial conditions with all of the following sets of propositions, that serve in effect as a simple type of declarative program, telling us all that we need to know about the anatomy and behavior of the truncated TM in question.

Mediate Conditions

\(\begin{array}{l} \texttt{(~p0\_q\#~(~p1\_q\#~))} \\ \texttt{(~p0\_q*~(~p1\_q*~))} \\ \\ \texttt{(~p1\_q\#~(~p2\_q\#~))} \\ \texttt{(~p1\_q*~(~p2\_q*~))} \end{array}\)

Terminal Conditions

\(\begin{array}{l} \texttt{((~p2\_q\#~)(~p2\_q*~))} \end{array}\)

State Partition

\(\begin{array}{l} \texttt{((~p0\_q0~),(~p0\_q1~),(~p0\_q\#~),(~p0\_q*~))} \\ \texttt{((~p1\_q0~),(~p1\_q1~),(~p1\_q\#~),(~p1\_q*~))} \\ \texttt{((~p2\_q0~),(~p2\_q1~),(~p2\_q\#~),(~p2\_q*~))} \end{array}\)

Register Partition

\(\begin{array}{l} \texttt{((~p0\_r0~),(~p0\_r1~),(~p0\_r2~))} \\ \texttt{((~p1\_r0~),(~p1\_r1~),(~p1\_r2~))} \\ \texttt{((~p2\_r0~),(~p2\_r1~),(~p2\_r2~))} \end{array}\)

Symbol Partition

\(\begin{array}{l} \texttt{((~p0\_r0\_s0~),(~p0\_r0\_s1~),(~p0\_r0\_s\#~))} \\ \texttt{((~p0\_r1\_s0~),(~p0\_r1\_s1~),(~p0\_r1\_s\#~))} \\ \texttt{((~p0\_r2\_s0~),(~p0\_r2\_s1~),(~p0\_r2\_s\#~))} \\ \\ \texttt{((~p1\_r0\_s0~),(~p1\_r0\_s1~),(~p1\_r0\_s\#~))} \\ \texttt{((~p1\_r1\_s0~),(~p1\_r1\_s1~),(~p1\_r1\_s\#~))} \\ \texttt{((~p1\_r2\_s0~),(~p1\_r2\_s1~),(~p1\_r2\_s\#~))} \\ \\ \texttt{((~p2\_r0\_s0~),(~p2\_r0\_s1~),(~p2\_r0\_s\#~))} \\ \texttt{((~p2\_r1\_s0~),(~p2\_r1\_s1~),(~p2\_r1\_s\#~))} \\ \texttt{((~p2\_r2\_s0~),(~p2\_r2\_s1~),(~p2\_r2\_s\#~))} \end{array}\)

Interaction Conditions

\(\begin{array}{l} \texttt{((~p0\_r0~) ~p0\_r0\_s0~ (~p1\_r0\_s0~))} \\ \texttt{((~p0\_r0~) ~p0\_r0\_s1~ (~p1\_r0\_s1~))} \\ \texttt{((~p0\_r0~) ~p0\_r0\_s\#~ (~p1\_r0\_s\#~))} \\ \\ \texttt{((~p0\_r1~) ~p0\_r1\_s0~ (~p1\_r1\_s0~))} \\ \texttt{((~p0\_r1~) ~p0\_r1\_s1~ (~p1\_r1\_s1~))} \\ \texttt{((~p0\_r1~) ~p0\_r1\_s\#~ (~p1\_r1\_s\#~))} \\ \\ \texttt{((~p0\_r2~) ~p0\_r2\_s0~ (~p1\_r2\_s0~))} \\ \texttt{((~p0\_r2~) ~p0\_r2\_s1~ (~p1\_r2\_s1~))} \\ \texttt{((~p0\_r2~) ~p0\_r2\_s\#~ (~p1\_r2\_s\#~))} \\ \\ \texttt{((~p1\_r0~) ~p1\_r0\_s0~ (~p2\_r0\_s0~))} \\ \texttt{((~p1\_r0~) ~p1\_r0\_s1~ (~p2\_r0\_s1~))} \\ \texttt{((~p1\_r0~) ~p1\_r0\_s\#~ (~p2\_r0\_s\#~))} \\ \\ \texttt{((~p1\_r1~) ~p1\_r1\_s0~ (~p2\_r1\_s0~))} \\ \texttt{((~p1\_r1~) ~p1\_r1\_s1~ (~p2\_r1\_s1~))} \\ \texttt{((~p1\_r1~) ~p1\_r1\_s\#~ (~p2\_r1\_s\#~))} \\ \\ \texttt{((~p1\_r2~) ~p1\_r2\_s0~ (~p2\_r2\_s0~))} \\ \texttt{((~p1\_r2~) ~p1\_r2\_s1~ (~p2\_r2\_s1~))} \\ \texttt{((~p1\_r2~) ~p1\_r2\_s\#~ (~p2\_r2\_s\#~))} \end{array}\)

Transition Relations

\(\begin{array}{l} \texttt{(~p0\_q0~~p0\_r1~~p0\_r1\_s0~~(~p1\_q0~~p1\_r2~~p1\_r1\_s0~))} \\ \texttt{(~p0\_q0~~p0\_r1~~p0\_r1\_s1~~(~p1\_q1~~p1\_r2~~p1\_r1\_s1~))} \\ \texttt{(~p0\_q0~~p0\_r1~~p0\_r1\_s\#~~(~p1\_q\#~~p1\_r0~~p1\_r1\_s\#~))} \\ \texttt{(~p0\_q0~~p0\_r2~~p0\_r2\_s\#~~(~p1\_q\#~~p1\_r1~~p1\_r2\_s\#~))} \\ \\ \texttt{(~p0\_q1~~p0\_r1~~p0\_r1\_s0~~(~p1\_q1~~p1\_r2~~p1\_r1\_s0~))} \\ \texttt{(~p0\_q1~~p0\_r1~~p0\_r1\_s1~~(~p1\_q0~~p1\_r2~~p1\_r1\_s1~))} \\ \texttt{(~p0\_q1~~p0\_r1~~p0\_r1\_s\#~~(~p1\_q*~~p1\_r0~~p1\_r1\_s\#~))} \\ \texttt{(~p0\_q1~~p0\_r2~~p0\_r2\_s\#~~(~p1\_q*~~p1\_r1~~p1\_r2\_s\#~))} \\ \\ \texttt{(~p1\_q0~~p1\_r1~~p1\_r1\_s0~~(~p2\_q0~~p2\_r2~~p2\_r1\_s0~))} \\ \texttt{(~p1\_q0~~p1\_r1~~p1\_r1\_s1~~(~p2\_q1~~p2\_r2~~p2\_r1\_s1~))} \\ \texttt{(~p1\_q0~~p1\_r1~~p1\_r1\_s\#~~(~p2\_q\#~~p2\_r0~~p2\_r1\_s\#~))} \\ \texttt{(~p1\_q0~~p1\_r2~~p1\_r2\_s\#~~(~p2\_q\#~~p2\_r1~~p2\_r2\_s\#~))} \\ \\ \texttt{(~p1\_q1~~p1\_r1~~p1\_r1\_s0~~(~p2\_q1~~p2\_r2~~p2\_r1\_s0~))} \\ \texttt{(~p1\_q1~~p1\_r1~~p1\_r1\_s1~~(~p2\_q0~~p2\_r2~~p2\_r1\_s1~))} \\ \texttt{(~p1\_q1~~p1\_r1~~p1\_r1\_s\#~~(~p2\_q*~~p2\_r0~~p2\_r1\_s\#~))} \\ \texttt{(~p1\_q1~~p1\_r2~~p1\_r2\_s\#~~(~p2\_q*~~p2\_r1~~p2\_r2\_s\#~))} \end{array}\)

Interpretation of the Propositional Program

Let us now run through the propositional specification of \(\mathrm{Stunt}(2),\) our truncated TM, and paraphrase what it says in ordinary language.

Mediate Conditions

\(\begin{array}{l} \texttt{(~p0\_q\#~(~p1\_q\#~))} \\ \texttt{(~p0\_q*~(~p1\_q*~))} \\ \\ \texttt{(~p1\_q\#~(~p2\_q\#~))} \\ \texttt{(~p1\_q*~(~p2\_q*~))} \end{array}\)

In the interpretation of the cactus language for propositional logic that we are using here, an expression of the form \(\texttt{(p(q))}\) expresses a conditional, an implication, or an if-then proposition, commonly read in one of the following ways:

\(\begin{array}{l} \mathrm{not}~ p ~\mathrm{without}~ q \\[4pt] p ~\mathrm{implies}~ q \\[4pt] \mathrm{if}~ p ~\mathrm{then}~ q \\[4pt] p \Rightarrow q \end{array}\)

A text string expression of the form \(\texttt{(p(q))}\) corresponds to a graph-theoretic data-structure of the following form:


o---------------------------------------o
|                                       |
|                 p   q                 |
|                 o---o                 |
|                 |                     |
|                 @                     |
|                                       |
o---------------------------------------o
|               ( p ( q ))              |
o---------------------------------------o


Taken together, the Mediate Conditions state the following:

If \(\mathrm{M}\) at \(p_0\!\) is in state \(q_\#,\!\) then \(\mathrm{M}\) at \(p_1\!\) is in state \(q_\#,\!\) and

If \(\mathrm{M}\) at \(p_0\!\) is in state \(q_*,\!\) then \(\mathrm{M}\) at \(p_1\!\) is in state \(q_*,\!\) and

If \(\mathrm{M}\) at \(p_1\!\) is in state \(q_\#,\!\) then \(\mathrm{M}\) at \(p_2\!\) is in state \(q_\#,\!\) and

If \(\mathrm{M}\) at \(p_1\!\) is in state \(q_*,\!\) then \(\mathrm{M}\) at \(p_2\!\) is in state \(q_*.\!\)

Terminal Conditions

\(\begin{array}{l} \texttt{((~p2\_q\#~)(~p2\_q*~))} \end{array}\)

In cactus syntax, an expression of the form \(\texttt{((p)(q))}\) expresses the disjunction \(p ~\mathrm{or}~ q.\) The corresponding cactus graph, here just a tree, has the following shape:


o---------------------------------------o
|                                       |
|                 p   q                 |
|                 o   o                 |
|                  \ /                  |
|                   o                   |
|                   |                   |
|                   @                   |
|                                       |
o---------------------------------------o
|               ((p) (q))               |
o---------------------------------------o


In effect, the Terminal Conditions state the following:

At time \(p_2\!\) machine \(\mathrm{M}\) is in state \(q_\#,\!\) or

At time \(p_2\!\) machine \(\mathrm{M}\) is in state \(q_*.\!\)

State Partition

\(\begin{array}{l} \texttt{((~p0\_q0~),(~p0\_q1~),(~p0\_q\#~),(~p0\_q*~))} \\ \texttt{((~p1\_q0~),(~p1\_q1~),(~p1\_q\#~),(~p1\_q*~))} \\ \texttt{((~p2\_q0~),(~p2\_q1~),(~p2\_q\#~),(~p2\_q*~))} \end{array}\)

In cactus syntax, an expression of the form \(\texttt{((} e_1 \texttt{),(} e_2 \texttt{),(} \ldots \texttt{),(} e_k \texttt{))}\!\) expresses a statement to the effect that exactly one of the expressions \(e_j\!\) is true, for \(j = 1 ~\mathit{to}~ k.\) Expressions of this form are called universal partition expressions, and the corresponding painted and rooted cactus (PARC) has the following shape:


o---------------------------------------o
|                                       |
|         e_1   e_2   ...   e_k         |
|          o     o           o          |
|          |     |           |          |
|          o-----o--- ... ---o          |
|           \               /           |
|            \             /            |
|             \           /             |
|              \         /              |
|               \       /               |
|                \     /                |
|                 \   /                 |
|                  \ /                  |
|                   @                   |
|                                       |
o---------------------------------------o
|       ((e_1),(e_2),(...),(e_k))       |
o---------------------------------------o


The State Partition segment of the propositional program consists of three universal partition expressions, taken in conjunction expressing the condition that \(\mathrm{M}\) has to be in one and only one of its states at each point in time under consideration. In short, we have the constraint:

At each of the points in time \(p_i,\!\) for \(i\!\) in the set \(\{ 0, 1, 2 \},\!\)

\(\mathrm{M}\) can be in exactly one state \(q_j,\!\) for \(j\!\) in the set \(\{ 0, 1, \#, * \}.\!\)

Register Partition

\(\begin{array}{l} \texttt{((~p0\_r0~),(~p0\_r1~),(~p0\_r2~))} \\ \texttt{((~p1\_r0~),(~p1\_r1~),(~p1\_r2~))} \\ \texttt{((~p2\_r0~),(~p2\_r1~),(~p2\_r2~))} \end{array}\)

The Register Partition segment of the propositional program consists of three universal partition expressions, taken in conjunction saying that the read head \(\mathrm{H}\) must be reading one and only one of the registers or tape cells available to it at each of the points in time under consideration. In sum:

At each of the points in time \(p_i,\!\) for \(i = 0, 1, 2,\!\)

\(\mathrm{H}\) is reading exactly one cell \(r_j,\!\) for \(j = 0, 1, 2.\!\)

Symbol Partition

\(\begin{array}{l} \texttt{((~p0\_r0\_s0~),(~p0\_r0\_s1~),(~p0\_r0\_s\#~))} \\ \texttt{((~p0\_r1\_s0~),(~p0\_r1\_s1~),(~p0\_r1\_s\#~))} \\ \texttt{((~p0\_r2\_s0~),(~p0\_r2\_s1~),(~p0\_r2\_s\#~))} \\ \\ \texttt{((~p1\_r0\_s0~),(~p1\_r0\_s1~),(~p1\_r0\_s\#~))} \\ \texttt{((~p1\_r1\_s0~),(~p1\_r1\_s1~),(~p1\_r1\_s\#~))} \\ \texttt{((~p1\_r2\_s0~),(~p1\_r2\_s1~),(~p1\_r2\_s\#~))} \\ \\ \texttt{((~p2\_r0\_s0~),(~p2\_r0\_s1~),(~p2\_r0\_s\#~))} \\ \texttt{((~p2\_r1\_s0~),(~p2\_r1\_s1~),(~p2\_r1\_s\#~))} \\ \texttt{((~p2\_r2\_s0~),(~p2\_r2\_s1~),(~p2\_r2\_s\#~))} \end{array}\)

The Symbol Partition segment of the propositional program for \(\mathrm{Stunt}(2)\) consists of nine universal partition expressions, taken in conjunction stipulating that there has to be one and only one symbol in each of the registers at each point in time under consideration. In short, we have:

At each of the points in time \(p_i,\!\) for \(i\!\) in \(\{ 0, 1, 2 \},\!\)

in each of the tape registers \(r_j,\!\) for \(j\!\) in \(\{ 0, 1, 2 \},\!\)

there can be exactly one sign \(s_k,\!\) for \(k\!\) in \(\{ 0, 1, \# \}.\!\)

Interaction Conditions

\(\begin{array}{l} \texttt{((~p0\_r0~) ~p0\_r0\_s0~ (~p1\_r0\_s0~))} \\ \texttt{((~p0\_r0~) ~p0\_r0\_s1~ (~p1\_r0\_s1~))} \\ \texttt{((~p0\_r0~) ~p0\_r0\_s\#~ (~p1\_r0\_s\#~))} \\ \\ \texttt{((~p0\_r1~) ~p0\_r1\_s0~ (~p1\_r1\_s0~))} \\ \texttt{((~p0\_r1~) ~p0\_r1\_s1~ (~p1\_r1\_s1~))} \\ \texttt{((~p0\_r1~) ~p0\_r1\_s\#~ (~p1\_r1\_s\#~))} \\ \\ \texttt{((~p0\_r2~) ~p0\_r2\_s0~ (~p1\_r2\_s0~))} \\ \texttt{((~p0\_r2~) ~p0\_r2\_s1~ (~p1\_r2\_s1~))} \\ \texttt{((~p0\_r2~) ~p0\_r2\_s\#~ (~p1\_r2\_s\#~))} \\ \\ \texttt{((~p1\_r0~) ~p1\_r0\_s0~ (~p2\_r0\_s0~))} \\ \texttt{((~p1\_r0~) ~p1\_r0\_s1~ (~p2\_r0\_s1~))} \\ \texttt{((~p1\_r0~) ~p1\_r0\_s\#~ (~p2\_r0\_s\#~))} \\ \\ \texttt{((~p1\_r1~) ~p1\_r1\_s0~ (~p2\_r1\_s0~))} \\ \texttt{((~p1\_r1~) ~p1\_r1\_s1~ (~p2\_r1\_s1~))} \\ \texttt{((~p1\_r1~) ~p1\_r1\_s\#~ (~p2\_r1\_s\#~))} \\ \\ \texttt{((~p1\_r2~) ~p1\_r2\_s0~ (~p2\_r2\_s0~))} \\ \texttt{((~p1\_r2~) ~p1\_r2\_s1~ (~p2\_r2\_s1~))} \\ \texttt{((~p1\_r2~) ~p1\_r2\_s\#~ (~p2\_r2\_s\#~))} \end{array}\)

In briefest terms, the Interaction Conditions simply express the circumstance that the mark on a tape cell cannot change between two points in time unless the tape head is over the cell in question at the initial one of those points in time. All that we have to do is to see how they manage to say this.

Consider a cactus expression of the following form:

\(\begin{array}{l} \texttt{((}~ p_i\_r_j ~\texttt{)}~ p_i\_r_j\_s_k ~\texttt{(}~ p_{i+1}\_r_j\_s_k ~\texttt{))} \end{array}\)

This expression has the corresponding cactus graph:

o---------------------------------------o
|                                       |
|         p<i>_r<j>   p<i+1>_r<j>_s<k>  |
|                 o   o                 |
|                  \ /                  |
|    p<i>_r<j>_s<k> o                   |
|                   |                   |
|                   @                   |
|                                       |
o---------------------------------------o

A propositional expression of this form can be read as follows:

\(\mathrm{If}\)
At the time \(p_i,\!\) the tape cell \(r_j\!\) bears the mark \(s_k,\!\)
\(\mathrm{But}\) it is not the case that:
At the time \(p_i,\!\) the tape head is on the tape cell \(r_j,\!\)
\(\mathrm{Then}\)
At the time \(p_{i+1},\!\) the tape cell \(r_j\!\) bears the mark \(s_k.\!\)

The eighteen clauses of the Interaction Conditions simply impose one such constraint on symbol changes for each combination of the times \(p_0, p_1,\!\) registers \(r_0, r_1, r_2,\!\) and symbols \(s_0, s_1, s_\#.\!\)

Transition Relations

\(\begin{array}{l} \texttt{(~p0\_q0~~p0\_r1~~p0\_r1\_s0~~(~p1\_q0~~p1\_r2~~p1\_r1\_s0~))} \\ \texttt{(~p0\_q0~~p0\_r1~~p0\_r1\_s1~~(~p1\_q1~~p1\_r2~~p1\_r1\_s1~))} \\ \texttt{(~p0\_q0~~p0\_r1~~p0\_r1\_s\#~~(~p1\_q\#~~p1\_r0~~p1\_r1\_s\#~))} \\ \texttt{(~p0\_q0~~p0\_r2~~p0\_r2\_s\#~~(~p1\_q\#~~p1\_r1~~p1\_r2\_s\#~))} \\ \\ \texttt{(~p0\_q1~~p0\_r1~~p0\_r1\_s0~~(~p1\_q1~~p1\_r2~~p1\_r1\_s0~))} \\ \texttt{(~p0\_q1~~p0\_r1~~p0\_r1\_s1~~(~p1\_q0~~p1\_r2~~p1\_r1\_s1~))} \\ \texttt{(~p0\_q1~~p0\_r1~~p0\_r1\_s\#~~(~p1\_q*~~p1\_r0~~p1\_r1\_s\#~))} \\ \texttt{(~p0\_q1~~p0\_r2~~p0\_r2\_s\#~~(~p1\_q*~~p1\_r1~~p1\_r2\_s\#~))} \\ \\ \texttt{(~p1\_q0~~p1\_r1~~p1\_r1\_s0~~(~p2\_q0~~p2\_r2~~p2\_r1\_s0~))} \\ \texttt{(~p1\_q0~~p1\_r1~~p1\_r1\_s1~~(~p2\_q1~~p2\_r2~~p2\_r1\_s1~))} \\ \texttt{(~p1\_q0~~p1\_r1~~p1\_r1\_s\#~~(~p2\_q\#~~p2\_r0~~p2\_r1\_s\#~))} \\ \texttt{(~p1\_q0~~p1\_r2~~p1\_r2\_s\#~~(~p2\_q\#~~p2\_r1~~p2\_r2\_s\#~))} \\ \\ \texttt{(~p1\_q1~~p1\_r1~~p1\_r1\_s0~~(~p2\_q1~~p2\_r2~~p2\_r1\_s0~))} \\ \texttt{(~p1\_q1~~p1\_r1~~p1\_r1\_s1~~(~p2\_q0~~p2\_r2~~p2\_r1\_s1~))} \\ \texttt{(~p1\_q1~~p1\_r1~~p1\_r1\_s\#~~(~p2\_q*~~p2\_r0~~p2\_r1\_s\#~))} \\ \texttt{(~p1\_q1~~p1\_r2~~p1\_r2\_s\#~~(~p2\_q*~~p2\_r1~~p2\_r2\_s\#~))} \end{array}\)

The Transition Relation segment of the propositional program for \(\mathrm{Stunt}(2)\) consists of sixteen implication statements with complex antecedents and consequents. Taken together, these give propositional expression to the TM Figure and Table that were given at the outset.

Just by way of a single example, consider the clause:

\(\texttt{(~p0\_q0~~p0\_r1~~p0\_r1\_s1~~(~p1\_q1~~p1\_r2~~p1\_r1\_s1~))}\)

This complex implication statement can be read to say:

\(\mathrm{If}\)
At the time \(p_0,\!\) the machine \(\mathrm{M}\) is in the state \(q_0,\!\) and
At the time \(p_0,\!\) the scanner \(\mathrm{H}\) is reading cell \(r_1,\!\) and
At the time \(p_0,\!\) the tape cell \(r_1\!\) contains a \(\texttt{1},\)
\(\mathrm{Then}\)
At the time \(p_1,\!\) the machine \(\mathrm{M}\) is in the state \(q_1,\!\) and
At the time \(p_1,\!\) the scanner \(\mathrm{H}\) is reading cell \(r_2,\!\) and
At the time \(p_1,\!\) the tape cell \(r_1\!\) contains a \(\texttt{1}.\)

Computation

The propositional program for \(\mathrm{Stunt}(2)\) uses the following set of \(9 + 12 + 36 = 57\!\) basic propositions or boolean variables:

\(\begin{matrix} \texttt{p0\_r0}, & \texttt{p0\_r1}, & \texttt{p0\_r2}, \\[6pt] \texttt{p1\_r0}, & \texttt{p1\_r1}, & \texttt{p1\_r2}, \\[6pt] \texttt{p2\_r0}, & \texttt{p2\_r1}, & \texttt{p2\_r2}. \end{matrix}\)

\(\begin{matrix} \texttt{p0\_q\#}, & \texttt{p0\_q*}, & \texttt{p0\_q0}, & \texttt{p0\_q1}, \\[6pt] \texttt{p1\_q\#}, & \texttt{p1\_q*}, & \texttt{p1\_q0}, & \texttt{p1\_q1}, \\[6pt] \texttt{p2\_q\#}, & \texttt{p2\_q*}, & \texttt{p2\_q0}, & \texttt{p2\_q1}. \end{matrix}\)

\(\begin{matrix} \texttt{p0\_r0\_s\#}, & \texttt{p0\_r0\_s*}, & \texttt{p0\_r0\_s0}, & \texttt{p0\_r0\_s1}, \\[4pt] \texttt{p0\_r1\_s\#}, & \texttt{p0\_r1\_s*}, & \texttt{p0\_r1\_s0}, & \texttt{p0\_r1\_s1}, \\[4pt] \texttt{p0\_r2\_s\#}, & \texttt{p0\_r2\_s*}, & \texttt{p0\_r2\_s0}, & \texttt{p0\_r2\_s1}, \\[12pt] \texttt{p1\_r0\_s\#}, & \texttt{p1\_r0\_s*}, & \texttt{p1\_r0\_s0}, & \texttt{p1\_r0\_s1}, \\[4pt] \texttt{p1\_r1\_s\#}, & \texttt{p1\_r1\_s*}, & \texttt{p1\_r1\_s0}, & \texttt{p1\_r1\_s1}, \\[4pt] \texttt{p1\_r2\_s\#}, & \texttt{p1\_r2\_s*}, & \texttt{p1\_r2\_s0}, & \texttt{p1\_r2\_s1}, \\[12pt] \texttt{p2\_r0\_s\#}, & \texttt{p2\_r0\_s*}, & \texttt{p2\_r0\_s0}, & \texttt{p2\_r0\_s1}, \\[4pt] \texttt{p2\_r1\_s\#}, & \texttt{p2\_r1\_s*}, & \texttt{p2\_r1\_s0}, & \texttt{p2\_r1\_s1}, \\[4pt] \texttt{p2\_r2\_s\#}, & \texttt{p2\_r2\_s*}, & \texttt{p2\_r2\_s0}, & \texttt{p2\_r2\_s1}. \end{matrix}\)

This means that the propositional program itself is nothing but a single proposition or boolean function of the form \(p : \mathbb{B}^{57} \to \mathbb{B}.\)

An assignment of boolean values to the above set of boolean variables is called an interpretation of the proposition \(p,\!\) and any interpretation of \(p\!\) that makes the proposition \(p : \mathbb{B}^{57} \to \mathbb{B}\) evaluate to \(1\!\) is referred to as a satisfying interpretation of the proposition \(p.\!\) Another way to specify interpretations, instead of giving them as bit vectors in \(\mathbb{B}^{57}\) and trying to remember some arbitrary ordering of variables, is to give them in the form of singular propositions, that is, a conjunction of the form \(e_1 \cdot \ldots \cdot e_{57}\) where each \(e_j\!\) is either \(v_j\!\) or \(\texttt{(} v_j \texttt{)},\) that is, either the assertion or the negation of the boolean variable \({v_j},\!\) as \(j\!\) runs from 1 to 57. Even more briefly, the same information can be communicated simply by giving the conjunction of the asserted variables, with the understanding that each of the others is negated.

A satisfying interpretation of the proposition \(p\!\) supplies us with all the information of a complete execution history for the corresponding program, and so all we have to do in order to get the output of the program \(p\!\) is to read off the proper part of the data from the expression of this interpretation.

Output

One component of the \(\begin{smallmatrix}\mathrm{Theme~One}\end{smallmatrix}\) program that I wrote some years ago finds all the satisfying interpretations of propositions expressed in cactus syntax. It's not a polynomial time algorithm, as you may guess, but it was just barely efficient enough to do this example in the 500 K of spare memory that I had on an old 286 PC in about 1989, so I will give you the actual outputs from those trials.

Output Conditions for Tape Input "0"

Let \(p_0\!\) be the proposition that we get by conjoining the proposition that describes the initial conditions for tape input "0" with the proposition that describes the truncated turing machine \(\mathrm{Stunt}(2).\) As it turns out, \(p_0\!\) has a single satisfying interpretation. This interpretation is expressible in the form of a singular proposition, which can in turn be indicated by its positive logical features, as shown in the following display:


o-------------------------------------------------o
|                                                 |
| p0_q0                                           |
|  p0_r1                                          |
|   p0_r0_s#                                      |
|    p0_r1_s0                                     |
|     p0_r2_s#                                    |
|      p1_q0                                      |
|       p1_r2                                     |
|        p1_r2_s#                                 |
|         p1_r0_s#                                |
|          p1_r1_s0                               |
|           p2_q#                                 |
|            p2_r1                                |
|             p2_r0_s#                            |
|              p2_r1_s0                           |
|               p2_r2_s#                          |
|                                                 |
o-------------------------------------------------o


The Output Conditions for Tape Input "0" can be read as follows:

At the time \(p_0,\!\) machine \(\mathrm{M}\) is in the state \(q_0,\!\) and

At the time \(p_0,\!\) scanner \(\mathrm{H}\) is reading cell \(r_1,\!\) and

At the time \(p_0,\!\) cell \(r_0\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_0,\!\) cell \(r_1\!\) contains the symbol \(\texttt{0},\) and

At the time \(p_0,\!\) cell \(r_2\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_1,\!\) machine \(\mathrm{M}\) is in the state \(q_0,\!\) and

At the time \(p_1,\!\) scanner \(\mathrm{H}\) is reading cell \(r_2,\!\) and

At the time \(p_1,\!\) cell \(r_0\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_1,\!\) cell \(r_1\!\) contains the symbol \(\texttt{0},\) and

At the time \(p_1,\!\) cell \(r_2\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_2,\!\) machine \(\mathrm{M}\) is in the state \(q_\#,\!\) and

At the time \(p_2,\!\) scanner \(\mathrm{H}\) is reading cell \(r_1,\!\) and

At the time \(p_2,\!\) cell \(r_0\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_2,\!\) cell \(r_1\!\) contains the symbol \(\texttt{0},\) and

At the time \(p_2,\!\) cell \(r_2\!\) contains the symbol \(\texttt{\#}.\)

The output of \(\mathrm{Stunt}(2)\) being the symbol that rests under the tape head \(\mathrm{H}\) if and when the machine \(\mathrm{M}\) reaches one of its resting states, we get the result that \(\mathrm{Parity}(0) = 0.\)

Output Conditions for Tape Input "1"

Let \(p_1\!\) be the proposition that we get by conjoining the proposition that describes the initial conditions for tape input "1" with the proposition that describes the truncated turing machine \(\mathrm{Stunt}(2).\) As it turns out, \(p_1\!\) has a single satisfying interpretation. This interpretation is expressible in the form of a singular proposition, which can in turn be indicated by its positive logical features, as shown in the following display:


o-------------------------------------------------o
|                                                 |
| p0_q0                                           |
|  p0_r1                                          |
|   p0_r0_s#                                      |
|    p0_r1_s1                                     |
|     p0_r2_s#                                    |
|      p1_q1                                      |
|       p1_r2                                     |
|        p1_r2_s#                                 |
|         p1_r0_s#                                |
|          p1_r1_s1                               |
|           p2_q*                                 |
|            p2_r1                                |
|             p2_r0_s#                            |
|              p2_r1_s1                           |
|               p2_r2_s#                          |
|                                                 |
o-------------------------------------------------o


The Output Conditions for Tape Input "1" can be read as follows:

At the time \(p_0,\!\) machine \(\mathrm{M}\) is in the state \(q_0,\!\) and

At the time \(p_0,\!\) scanner \(\mathrm{H}\) is reading cell \(r_1,\!\) and

At the time \(p_0,\!\) cell \(r_0\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_0,\!\) cell \(r_1\!\) contains the symbol \(\texttt{1},\) and

At the time \(p_0,\!\) cell \(r_2\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_1,\!\) machine \(\mathrm{M}\) is in the state \(q_1,\!\) and

At the time \(p_1,\!\) scanner \(\mathrm{H}\) is reading cell \(r_2,\!\) and

At the time \(p_1,\!\) cell \(r_0\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_1,\!\) cell \(r_1\!\) contains the symbol \(\texttt{1},\) and

At the time \(p_1,\!\) cell \(r_2\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_2,\!\) machine \(\mathrm{M}\) is in the state \(q_*,\!\) and

At the time \(p_2,\!\) scanner \(\mathrm{H}\) is reading cell \(r_1,\!\) and

At the time \(p_2,\!\) cell \(r_0\!\) contains the symbol \(\texttt{\#},\) and

At the time \(p_2,\!\) cell \(r_1\!\) contains the symbol \(\texttt{1},\) and

At the time \(p_2,\!\) cell \(r_2\!\) contains the symbol \(\texttt{\#}.\)

The output of \(\mathrm{Stunt}(2)\) being the symbol that rests under the tape head \(\mathrm{H}\) when and if the machine \(\mathrm{M}\) reaches one of its resting states, we get the result that \(\mathrm{Parity}(1) = 1.\)

Document History

Ontology List : Feb–Mar 2004

  1. http://suo.ieee.org/ontology/msg05457.html
  2. http://suo.ieee.org/ontology/msg05458.html
  3. http://suo.ieee.org/ontology/msg05459.html
  4. http://suo.ieee.org/ontology/msg05460.html
  5. http://suo.ieee.org/ontology/msg05461.html
  6. http://suo.ieee.org/ontology/msg05462.html
  7. http://suo.ieee.org/ontology/msg05463.html
  8. http://suo.ieee.org/ontology/msg05464.html
  9. http://suo.ieee.org/ontology/msg05465.html
  10. http://suo.ieee.org/ontology/msg05466.html
  11. http://suo.ieee.org/ontology/msg05467.html
  12. http://suo.ieee.org/ontology/msg05469.html
  13. http://suo.ieee.org/ontology/msg05470.html
  14. http://suo.ieee.org/ontology/msg05471.html
  15. http://suo.ieee.org/ontology/msg05472.html
  16. http://suo.ieee.org/ontology/msg05473.html
  17. http://suo.ieee.org/ontology/msg05474.html
  18. http://suo.ieee.org/ontology/msg05475.html
  19. http://suo.ieee.org/ontology/msg05476.html
  20. http://suo.ieee.org/ontology/msg05479.html

NKS Forum : Feb–Jun 2004

  1. http://forum.wolframscience.com/showthread.php?postid=664#post664
  2. http://forum.wolframscience.com/showthread.php?postid=666#post666
  3. http://forum.wolframscience.com/showthread.php?postid=677#post677
  4. http://forum.wolframscience.com/showthread.php?postid=684#post684
  5. http://forum.wolframscience.com/showthread.php?postid=689#post689
  6. http://forum.wolframscience.com/showthread.php?postid=697#post697
  7. http://forum.wolframscience.com/showthread.php?postid=708#post708
  8. http://forum.wolframscience.com/showthread.php?postid=721#post721
  9. http://forum.wolframscience.com/showthread.php?postid=722#post722
  10. http://forum.wolframscience.com/showthread.php?postid=725#post725
  11. http://forum.wolframscience.com/showthread.php?postid=733#post733
  12. http://forum.wolframscience.com/showthread.php?postid=756#post756
  13. http://forum.wolframscience.com/showthread.php?postid=759#post759
  14. http://forum.wolframscience.com/showthread.php?postid=764#post764
  15. http://forum.wolframscience.com/showthread.php?postid=766#post766
  16. http://forum.wolframscience.com/showthread.php?postid=767#post767
  17. http://forum.wolframscience.com/showthread.php?postid=773#post773
  18. http://forum.wolframscience.com/showthread.php?postid=775#post775
  19. http://forum.wolframscience.com/showthread.php?postid=777#post777
  20. http://forum.wolframscience.com/showthread.php?postid=791#post791
  21. http://forum.wolframscience.com/showthread.php?postid=1458#post1458
  22. http://forum.wolframscience.com/showthread.php?postid=1461#post1461
  23. http://forum.wolframscience.com/showthread.php?postid=1463#post1463
  24. http://forum.wolframscience.com/showthread.php?postid=1464#post1464
  25. http://forum.wolframscience.com/showthread.php?postid=1467#post1467
  26. http://forum.wolframscience.com/showthread.php?postid=1469#post1469
  27. http://forum.wolframscience.com/showthread.php?postid=1470#post1470
  28. http://forum.wolframscience.com/showthread.php?postid=1471#post1471
  29. http://forum.wolframscience.com/showthread.php?postid=1473#post1473
  30. http://forum.wolframscience.com/showthread.php?postid=1475#post1475
  31. http://forum.wolframscience.com/showthread.php?postid=1479#post1479
  32. http://forum.wolframscience.com/showthread.php?postid=1489#post1489
  33. http://forum.wolframscience.com/showthread.php?postid=1490#post1490

Inquiry List : Feb–Jun 2004

  1. http://stderr.org/pipermail/inquiry/2004-February/001228.html
  2. http://stderr.org/pipermail/inquiry/2004-February/001230.html
  3. http://stderr.org/pipermail/inquiry/2004-February/001231.html
  4. http://stderr.org/pipermail/inquiry/2004-February/001232.html
  5. http://stderr.org/pipermail/inquiry/2004-February/001233.html
  6. http://stderr.org/pipermail/inquiry/2004-February/001234.html
  7. http://stderr.org/pipermail/inquiry/2004-March/001235.html
  8. http://stderr.org/pipermail/inquiry/2004-March/001236.html
  9. http://stderr.org/pipermail/inquiry/2004-March/001237.html
  10. http://stderr.org/pipermail/inquiry/2004-March/001238.html
  11. http://stderr.org/pipermail/inquiry/2004-March/001240.html
  12. http://stderr.org/pipermail/inquiry/2004-March/001242.html
  13. http://stderr.org/pipermail/inquiry/2004-March/001243.html
  14. http://stderr.org/pipermail/inquiry/2004-March/001244.html
  15. http://stderr.org/pipermail/inquiry/2004-March/001245.html
  16. http://stderr.org/pipermail/inquiry/2004-March/001246.html
  17. http://stderr.org/pipermail/inquiry/2004-March/001247.html
  18. http://stderr.org/pipermail/inquiry/2004-March/001248.html
  19. http://stderr.org/pipermail/inquiry/2004-March/001249.html
  20. http://stderr.org/pipermail/inquiry/2004-March/001255.html
  21. http://stderr.org/pipermail/inquiry/2004-June/001630.html
  22. http://stderr.org/pipermail/inquiry/2004-June/001631.html
  23. http://stderr.org/pipermail/inquiry/2004-June/001632.html
  24. http://stderr.org/pipermail/inquiry/2004-June/001633.html
  25. http://stderr.org/pipermail/inquiry/2004-June/001634.html
  26. http://stderr.org/pipermail/inquiry/2004-June/001635.html
  27. http://stderr.org/pipermail/inquiry/2004-June/001636.html
  28. http://stderr.org/pipermail/inquiry/2004-June/001637.html
  29. http://stderr.org/pipermail/inquiry/2004-June/001638.html
  30. http://stderr.org/pipermail/inquiry/2004-June/001639.html
  31. http://stderr.org/pipermail/inquiry/2004-June/001640.html
  32. http://stderr.org/pipermail/inquiry/2004-June/001641.html
  33. http://stderr.org/pipermail/inquiry/2004-June/001642.html