SELF-ORGANIZING SYSTEMS
1963
Edited By
JAMES EMMETT GARVEY
Office of Naval Research
Pasadena, California
ACR-96
OFFICE OF NAVAL RESEARCH
DEPARTMENT OF THE NAVY
WASHINGTON, D.C.
For sale by the Superintendent of Documents.
U.S. Government Printing Office
Washington, D.C., 20402—Price $1.50
CONTENTS
| Foreword | [iv] |
| The Ionic Hypothesis and Neuron Models | [ 1] |
| —E. R. Lewis | |
| Fields and Waves in Excitable Cellular Structures | [19] |
| —R. M. Stewart | |
| Multi-Layer Learning Networks | [37] |
| —R. A. Stafford | |
| Adaptive Detection of Unknown Binary Waveforms | [46] |
| —J. J. Spilker, Jr. | |
| Conceptual Design of Self-Organizing Machines | [52] |
| —P. A. Kleyn | |
| A Topological Foundation for Self-Organization | [65] |
| —R. I. Ścibor-Marchocki | |
| On Functional Neuron Modeling | [71] |
| —C. E. Hendrix | |
| Selection of Parameters for Neural Net Simulations | [76] |
| —R. K. Overton | |
| Index of Invited Participants | [77] |
FOREWORD
The papers appearing in this volume were presented at a Symposium on Self-Organizing Systems, which was sponsored by the Office of Naval Research and held at the California Institute of Technology, Pasadena, California, on 14 November 1963. The Symposium was organized with the aim of providing a critical forum for the presentation and discussion of contemporary significant research efforts, with the emphasis on relatively uncommon approaches and methods in an early state of development. This aim and nature dictated that the Symposium be in effect a Working Group, with numerically limited invitational participation.
The papers which were presented and discussed did in fact serve to introduce several relatively unknown approaches; some of the speakers were promising young scientists, others had become known for contributions in different fields and were as yet unrecognized for their recent work in self-organization. In addition, the papers as a collection provided a particularly broad, cross-disciplinary spectrum of investigations which possessed intrinsic value as a portrayal of the bases upon which this new discipline rests. Accordingly, it became obvious in retrospect that the information presented and discussed at the Symposium was of considerable interest—and should thus receive commensurate dissemination—to a much broader group of scientists and engineers than those who were able to participate directly in the meeting itself. This volume is the result of that observation; as an edited collection of the papers presented at the Symposium, it forms the Proceedings thereof. If it provides a useful reference for present and future investigators, as well as documenting the source of several new approaches, it will have fulfilled its intended purpose well.
A Symposium which takes the nature of a Working Group depends for its utility especially upon effective commentary and critical analysis, and we commend all the participants for their contributions in this regard. It is appropriate, further, to acknowledge the contributions to the success of the Symposium made by the following: The California Institute of Technology for volunteering to act as host and for numerous supporting services; Professor Gilbert D. McCann, Director of the Willis Booth Computing Center at the California Institute of Technology, and the members of the technical and secretarial staffs of the Computing Center, who assumed the responsibility of acting as the immediate representatives of the Institute; the members of the Program Committee, who organized and led the separate sessions—Harold Hamilton of General Precision, Joseph Hawkins of Ford Motor Company, Robert Stewart of Space-General, Peter Kleyn of Northrop, and Professor McCann; members of the Technical Information Division of the Naval Research Laboratory, who published these Proceedings; and especially the authors of the papers, which comprised the heart of the Symposium and subsequently formed this volume. To all of these the sponsors wish to express their very sincere appreciation.
James Emmett Garvey
Office of Naval Research Branch Office
Pasadena, California
Margo A. Sass
Office of Naval Research
Washington, D.C.
The Ionic Hypothesis and Neuron Models
E. R. Lewis
Librascope Group, General Precision, Inc.
Research and Systems Center
Glendale, California
The measurements of Hodgkin and Huxley were aimed at revealing the mechanism of generation and propagation of the all-or-none spike. Their results led to the Modern Ionic Hypothesis. Since the publication of their papers in 1952, advanced techniques with microelectrodes have led to the discovery of many modes of subthreshold activity not only in the axon but also in the somata and dendrites of neurons. This activity includes synaptic potentials, local response potentials, and pacemaker potentials.
We considered the question, “Can this activity also be explained in terms of the Hodgkin-Huxley Model?” To seek an answer, we have constructed an electronic analog based on the ionic hypothesis and designed around the data of Hodgkin and Huxley. Synaptic inputs were simulated by simple first-order or second-order networks connected directly to simulated conductances (potassium or sodium). The analog has, with slight parameter adjustments, produced all modes of threshold and subthreshold activity.
INTRODUCTION
In recent years physiologists have become quite adept at probing into neurons with intracellular microelectrodes. They are now able, in fact, to measure (a) the voltage change across the postsynaptic membrane elicited by a single presynaptic impulse (see, for examples, references [1] and [2]) and (b) the voltage-current characteristics across a localized region of the nerve cell membrane [(3)], [(4)], [(5)], [(6)]. With microelectrodes, physiologists have been able to examine not only the all-or-none spike generating and propagating properties of axons but also the electrical properties of somatic and dendritic structures in individual neurons. The resulting observations have led many physiologists to believe that the individual nerve cell is a potentially complex information-processing system far removed from the simple two-state device envisioned by many early modelers. This new concept of the neuron is well summarized by Bullock in his 1959 Science article [(10)]. In the light of recent physiological literature, one cannot justifiably omit the diverse forms of somatic and dendritic behavior when assessing the information-processing capabilities of single neurons. This is true regardless of the means of assessment—whether one uses mathematical idealizations, electrochemical models, or electronic analogs. We have been interested specifically in electronic analogs of the neuron; and in view of the widely diversified behavior which we must simulate, our first goal has been to find a unifying concept about which to design our analogs. We believe we have found such a concept in the Modern Ionic Hypothesis, and in this paper we will discuss an electronic analog of the neuron which was based on this hypothesis and which simulated not only the properties of the axon but also the various subthreshold properties of the somata and dendrites of neurons.
We begin with a brief summary of the various types of subthreshold activity which have been observed in the somatic and dendritic structures of neurons. This is followed by a brief discussion of the Hodgkin-Huxley data and of the Modern Ionic Hypothesis. An electronic analog based on the Hodgkin-Huxley data is then introduced, and we show how this analog can be used to provide all of the various types of somatic and dendritic activity.
SUBTHRESHOLD ELECTRICAL ACTIVITY
IN NEURONS
In studying the recent literature in neurophysiology, one is immediately struck by the diversity in form of both elicited and spontaneous electrical activity in the single nerve cell. This applies not only to the temporal patterns of all-or-none action potentials but also to the graded somatic and dendritic potentials. The synaptic membrane of a neuron, for example, is often found to be electrically inexcitable and thus incapable of producing an action potential; yet the graded, synaptically induced potentials show an amazing diversity in form. In response to a presynaptic impulse, the postsynaptic membrane may become hyperpolarized (inhibitory postsynaptic potential), depolarized (excitatory postsynaptic potential), or remain at the resting potential but with an increased permeability to certain ions (a form of inhibition). The form of the postsynaptic potential in response to an isolated presynaptic spike may vary from synapse to synapse in several ways, as shown in [Figure 1]. Following a presynaptic spike, the postsynaptic potential typically rises with some delay to a peak value and then falls back toward the equilibrium or resting potential. Three potentially important factors are the delay time (synaptic delay), the peak amplitude (spatial weighting of synapse), and the rate of fall toward the equilibrium potential (temporal weighting of synapse). The responses of a synapse to individual spikes in a volley may be progressively enhanced (facilitation), diminished (antifacilitation), or neither [(1)], [(2)], [(7)], [(8)]. Facilitation may be in the form of progressively increased peak amplitude, or in the form of progressively decreased rate of fall ([see Figure 2]). The time course and magnitude of facilitation or antifacilitation may very well be important synaptic parameters. In addition, the postsynaptic membrane sometimes exhibits excitatory or inhibitory aftereffects (or both) on cessation of a volley of presynaptic spikes [(2)], [(7)]; and the time course and magnitude of the aftereffects may be important parameters. Clearly, even if one considers the synaptic potentials alone, he is faced with an impressive variety of responses. Examples of the various types of postsynaptic responses may be found in the literature, but for purposes of the present discussion the idealized wave forms in [Figure 2] will demonstrate the diversity of electrical behavior with which one is faced.
Figure 1—Excitatory postsynaptic potentials in response to a single presynaptic spike
Figure 2—Idealized postsynaptic potentials
In addition to synaptically induced potentials, low-frequency, spontaneous potential fluctuations have been observed in many neurons [(2)], [(7)], [(9)], [(10)], [(11)]. These fluctuations, generally referred to as pacemaker potentials, are usually rhythmic and may be undulatory or more nearly saw-toothed in form. The depolarizing phase may be accompanied by a spike, a volley of spikes, or no spikes at all. Pacemaker frequencies have been noted from ten or more cycles per second down to one cycle every ten seconds or more. Some idealized pacemaker wave forms are shown in [Figure 3].
Figure 3—Idealized pacemaker potentials
Figure 4—Graded response
Bullock [(7)], [(10)], [(12)], [(13)] has demonstrated the existence of a third type of subthreshold response, which he calls the graded response. While the postsynaptic membrane is quite often electrically inexcitable, other regions of the somatic and dendritic membranes appear to be moderately excitable. It is in these regions that Bullock observes the graded response. If one applies a series of pulsed voltage stimuli to the graded-response region, the observed responses would be similar to those shown in [Figure 4A]. Plotting the peak response voltage as a function of the stimulus voltage would result in a curve similar to that in [Figure 4B] ([see Ref. 3, page 4]). For small values of input voltage, the response curve is linear; the membrane is passive. As the stimulus voltage is increased, however, the response becomes more and more disproportionate. The membrane is actively amplifying the stimulus potential. At even higher values of stimulus potential, the system becomes regenerative; and a full action potential results. The peak amplitude of the response depends on the duration of the stimulus as well as on the amplitude. It also depends on the rate of application of the stimulus voltage. If the stimulus potential is a voltage ramp, for example, the response will depend on the slope of the ramp. If the rate of rise is sufficiently low, the membrane will respond in a passive manner to voltages much greater than the spike threshold for suddenly applied voltages. In other words, the graded-response regions appear to accommodate to slowly varying potentials.
In terms of functional operation, we can think of the synapse as a transducer. The input to this transducer is a spike or series of spikes in the presynaptic axon. The output is an accumulative, long-lasting potential which in some way (perhaps not uniquely) represents the pattern of presynaptic spikes. The pacemaker appears to perform the function of a clock, producing periodic spikes or spike bursts or producing periodic changes in the over-all excitability of the neuron. The graded-response regions appear to act as nonlinear amplifiers and, occasionally, spike initiators. The net result of this electrical activity is transformed into a series of spikes which originate at spike initiation sites and are propagated along axons to other neurons. The electrical activity in the neuron described above is summarized in the following outline (taken in part from [Bullock (7)]):
- 1. Synaptic Potentials
- a. Excitatory or inhibitory
- b. Facilitated, antifacilitated, or neither
- c. With excitatory aftereffect, inhibitory
- aftereffect, neither, or both
- 2. Pacemaker Potentials
- a. Relaxation type, undulatory type,
- or none at all
- b. Producing single spike, spike burst,
- or no spikes
- c. Rhythmic or sporadic
- 3. Graded Response (rate sensitive)
- 4. Spike Initiation
THE MODERN IONIC HYPOTHESIS
Hodgkin, Huxley, and Katz [(3)] and Hodgkin and Huxley [(14)], [(15)], [(16)], in 1952, published a series of papers describing detailed measurements of voltage, current, and time relationships in the giant axon of the squid (Loligo). Hodgkin and Huxley [(17)] consolidated and formalized these data into a set of simultaneous differential equations describing the hypothetical time course of events during spike generation and propagation. The hypothetical system which these equations describe is the basis of the Modern Ionic Hypothesis.
The system proposed by Hodgkin and Huxley is basically one of dynamic opposition of ionic fluxes across the axon membrane. The membrane itself forms the boundary between two liquid phases—the intracellular fluid and the extracellular fluid. The intracellular fluid is rich in potassium ions and immobile organic anions, while the extracellular fluid contains an abundance of sodium ions and chloride ions. The membrane is slightly permeable to the potassium, sodium, and chloride ions; so these ions tend to diffuse across the membrane. When the axon is inactive (not propagating a spike), the membrane is much more permeable to chloride and potassium ions than it is to sodium ions. In this state, in fact, sodium ions are actively transported from the inside of the membrane to the outside at a rate just sufficient to balance the inward leakage. The relative sodium ion concentrations on both sides of the membrane are thus fixed by the active transport rate, and the net sodium flux across the membrane is effectively zero. The potassium ions, on the other hand, tend to move out of the cell; while chloride ions tend to move into it. The inside of the cell thus becomes negative with respect to the outside. When the potential across the membrane is sufficient to balance the inward diffusion of chloride with an equal outward drift, and the outward diffusion of potassium with an inward drift (and possibly an inward active exchange), equilibrium is established. The equilibrium potential is normally in the range of 60 to 65 millivolts.
The resting neural membrane is thus polarized, with the inside approximately 60 millivolts negative with respect to the outside. Most of the Hodgkin-Huxley data is based on measurements of the transmembrane current in response to an imposed stepwise reduction (depolarization) of membrane potential. By varying the external ion concentrations, Hodgkin and Huxley were able to resolve the transmembrane current into two “active” components, the potassium ion current and the sodium ion current. They found that while the membrane permeabilities to chloride and most other inorganic ions were relatively constant, the permeabilities to both potassium and sodium were strongly dependent on membrane potential. In response to a suddenly applied (step) depolarization, the sodium permeability rises rapidly to a peak and then declines exponentially to a steady value. The potassium permeability, on the other hand, rises with considerable delay to a value which is maintained as long as the membrane remains depolarized. The magnitudes of both the potassium and the sodium permeabilities increase monotonically with increasing depolarization. A small imposed depolarization will result in an immediately increased sodium permeability. The resulting increased influx of sodium ions results in further depolarization; and the process becomes regenerative, producing the all-or-none action potential. At the peak of the action potential, the sodium conductance begins to decline, while the delayed potassium conductance is increasing. Recovery is brought about by an efflux of potassium ions, and both ionic permeabilities fall rapidly as the membrane is repolarized. The potassium permeability, however, falls less rapidly than that of sodium. This is basically the explanation of the all-or-none spike according to the Modern Ionic Hypothesis.
Figure 5—Hodgkin-Huxley representation of small area of axon membrane
Figure 6—Typical responses of sodium conductance and potassium conductance to imposed step depolarization
By defining the net driving force on any given ion species as the difference between the membrane potential and the equilibrium potential for that ion and describing permeability changes in terms of equivalent electrical conductance changes, Hodgkin and Huxley reduced the ionic model to the electrical equivalent in [Figure 5]. The important dynamic variables in this equivalent network are the sodium conductance (G{Na}) and the potassium conductance (G{K}). The change in the sodium conductance in response to a step depolarization is shown in [Figure 6B]. This change can be characterized by seven voltage dependent parameters:
1. Delay time—generally much less than 1 msec
2. Rise time—1 msec or less
3. Magnitude of peak conductance—increases monotonically with increasing depolarization
4. Inactivation time constant—decreases monotonically with increasing depolarization.
5. Time constant of recovery from inactivation—incomplete data
6. Magnitude of steady-state conductance—increases monotonically with increasing depolarization
7. Fall time on sudden repolarization—less than 1 msec.
[Figure 6B] shows the potassium conductance change in response to an imposed step depolarization. Four parameters are sufficient to characterize this response:
1. Delay time—decreases monotonically with increasing depolarization
2. Rise time—decreases monotonically with increasing depolarization
3. Magnitude of steady-state conductance—increases monotonically with increasing depolarization
4. Fall time on sudden repolarization—8 msec or more, decreases slightly with increasing depolarization.
In addition to the aforementioned parameters, the transient portion of the sodium conductance appears to exhibit an accommodation to slowly varying membrane potentials. The time constants of accommodation appear to be those of inactivation or recovery from inactivation—depending on the direction of change in the membrane potential [(18)]. The remaining elements in the Hodgkin-Huxley model are constant and are listed below:
- 1. Potassium potential—80 to 85 mv (inside negative)
- 2. Sodium potential—45 to 50 mv (inside positive)
- 3. Leakage potential—38 to 43 mv (inside negative)
- 4. Leakage conductance—approx. 0.23 millimhos/cm²
- 5. Membrane capacitance—approx. 1 μf/cm²
- 6. Resting potential—60 to 65 mv
- 7. Spike amplitude—approx. 100 mv
ELECTRONIC SIMULATION OF THE HODGKIN-HUXLEY MODEL
Figure 7—System diagram for electronic simulation of the Hodgkin-Huxley model
Given a suitable means of generating the conductance functions, GNa(v,t) and GK(v,t), one can readily stimulate the essential aspects of the Modern Ionic Hypothesis. If we wish to do this electronically, we have two problems. First, we must synthesize a network whose input is the membrane potential and whose output is a voltage or current proportional to the desired conductance function. Second, we must transform the output from a voltage or current to an effective electronic conductance. The former implies the need for nonlinear, active filters, while the latter implies the need for multipliers. The basic block diagram is shown in [Figure 7]. Several distinct realizations of this system have been developed in our laboratory, and in each case the results were the same. With parameters adjusted to closely match the data of Hodgkin and Huxley, the electronic model exhibits all of the important properties of the axon. It produces spikes of 1 to 2 msec duration with a threshold of approximately 5% to 10% of the spike amplitude. The applied stimulus is generally followed by a prepotential, then an active rise of less than 1 msec, followed by an active recovery. The after-depolarization generally lasts several msec, followed by a prolonged after-hyperpolarization. The model exhibits the typical strength-duration curve, with rheobase of 5% to 10% of the spike amplitude. For sufficiently prolonged sodium inactivation (long time constant of recovery from inactivation), the model also exhibits an effect identical to classical Wedensky inhibition [(18)]. Thus, as would be expected, the electronic model simulates very well the electrical properties of the axon.
In addition to the axon properties, however, the electronic model is able to reproduce all of the somatic and dendritic activity outlined in the section on subthreshold activity. Simulation of the pacemaker and graded-response potentials is accomplished without additional circuitry. In the case of synaptically induced potentials, however, auxiliary networks are required. These networks provide additive terms to the variable conductances in accordance with current notions on synaptic transmission [(19)]. Two types of networks have been used. In both, the inputs are simulated presynaptic spikes, and in both the outputs are the resulting simulated chemical transmitter concentration. In both, the transmitter substance was assumed to be injected at a constant rate during a presynaptic spike and subsequently inactivated in the presence of an enzyme. One network simulates a first-order chemical reaction, where the enzyme concentration is effectively constant. The other simulates a second-order chemical reaction, where the enzyme concentration is assumed to be reduced during the inactivation process. For simulation of an excitatory synapse, the output of the auxiliary network is added directly to GNa in the electronic model. For inhibition, it is added to GK. With the parameters of the electronic membrane model set at the values measured by Hodgkin and Huxley, we have attempted to simulate synaptic activity with the aid of the two types of auxiliary networks. In the case of the simulated first-order reaction, the excitatory synapse exhibits facilitation, antifacilitation, or neither—depending on the setting of a single parameter, the transmitter inactivation rate (i.e., the effective enzyme concentration). This parameter would appear, in passing, to be one of the most probable synaptic variables. In this case, the mechanisms for facilitation and antifacilitation are contained in the simulated postsynaptic membrane. Facilitation is due to the nonlinear dependence of GNa on membrane potential, while antifacilitation is due to inactivation of GNa. The occurrence of one form of response or the other is determined by the relative importance of the two mechanisms [(18)]. Grundfest [(20)] has mentioned both of these mechanisms as potentially facilitory and antifacilitory, respectively. The simulated inhibitory synapse with the first order input is capable of facilitation [(18)], but no antifacilitation has been observed. Again, the presence or absence of facilitation is determined by the inactivation rate.
With the simulated second-order reaction, both excitatory and inhibitory synapses exhibit facilitation. In this case, two facilitory mechanisms are present—one in the postsynaptic membrane and one in the nonconstant transmitter inactivation reaction. The active membrane currents can, in fact, be removed; and this system will still exhibit facilitation. With the second-order auxiliary network, the presence of excitatory facilitation, antifacilitation, or neither depends on the initial, or resting, transmitter inactivation rate. The synaptic behavior also depends parametrically on the simulated enzyme reactivation rate. Inhibitory antifacilitation can be introduced with either type of auxiliary network by limiting the simulated presynaptic transmitter supply.
Certain classes of aftereffects are inherent in the mechanisms of the Ionic Hypothesis. In the electronic model, aftereffects are observed following presynaptic volleys with either type of auxiliary network. Following a volley of spikes into the simulated excitatory synapse, for example, rebound hyperpolarization may or may not occur depending on the simulated transmitter inactivation rate. If the inactivation rate is sufficiently high, rebound will occur. This rebound can be monophasic (inhibitory phase only) or polyphasic (successive cycles of excitation and inhibition). Following a volley of spikes into the simulated inhibitory synapse, rebound depolarization may or may not occur depending on the simulated transmitter inactivation rate. This rebound can also be monophasic or polyphasic. Sustained postexcitatory depolarization and sustained postinhibitory hyperpolarization [(2)] have been achieved in the model by making the transmitter inactivation rate sufficiently low.
The general forms of the postsynaptic potentials simulated with the electronic model are strikingly similar to those published in the literature for real neurons. The first-order auxiliary network produces facilitation of a form almost identical to that shown by Otani and Bullock [(8)] while the second-order auxiliary network produces facilitation of the type shown by Chalazonitis and Arvanitake [(2)]. The excitatory antifacilitation is almost identical to that shown by Hagiwara and Bullock [(1)] in both form and dependence on presynaptic spike frequency. In every case, the synaptic behavior is determined by the effective rate of transmitter inactivation, which in real neurons would presumably be directly proportional to the effective concentration of inactivating enzyme at the synapse.
Pacemaker potentials are easily simulated with the electronic model without the use of auxiliary networks. This is achieved either by inserting a large, variable shunt resistor across the simulated membrane ([see Figure 5]) or by allowing a small sodium current leakage at the resting potential. With the remaining parameters of the model set as close as possible to the values determined by Hodgkin and Huxley, the leakage current induces low-frequency, spontaneous spiking. The spike frequency increases monotonically with increasing leakage current. In addition, if the sodium conductance inactivation is allowed to accumulate over several spikes, periodic spike pairs and spike bursts will result. Subthreshold pacemaker potentials have also been observed in the model, but with parameter values set close to the Hodgkin-Huxley data these are generally higher in frequency than pacemaker potentials in real neurons. It is interesting that a pacemaker mode may exist in the absence of the simulated sodium conductance. It is a very high-frequency mode (50 cps or more) and results from the alternating dominance of potassium current and chloride (or leakage ion) current in determining the membrane potential. The significance of this mode cannot be assessed until better data is available for the potassium conductance at low levels of depolarization in real neurons. In general, as far as the model is concerned, pacemaker potentials are possible because the potassium conductance is delayed in both its rise with depolarization and its fall with repolarization.
Rate sensitive graded response has also been observed in the electronic model. The rate sensitivity—or accommodation—is due to the sodium conductance inactivation. The response of the model to an imposed ramp depolarization was discussed in [Reference 18]. At this time, several alternative model parameters could be altered to bring about reduced electrical excitability. None of the parameter changes was very satisfying, however, because none of them was in any way justified by physiological data. We have since found that the membrane capacitance, a plausible parameter in view of recent physiological findings, can completely determine the electrical excitability. Thus, with the capacitance determined by Hodgkin and Huxley (1 microfarad per cm²), the model exhibits excitability characteristic of the axon. As the capacitance is increased, the model becomes less excitable until, with 10 or 12 μμf, it is effectively inexcitable. Thus, with an increased capacitance—but with all the remaining parameters set as close as possible to the Hodgkin-Huxley values—the electronic model exhibits the characteristics of Bullock’s graded-response regions.
Whether membrane capacitance is the determining factor in real neurons is, of course, a matter of speculation. Quite a controversy is raging over membrane capacity measurements ([see Rall (21)]), but the evidence indicates that the capacity in the soma is considerably greater than that in the axon [(6)], [(22)].
It should be added that increasing the capacitance until the membrane model becomes inexcitable has little effect on the variety of available simulated synaptic responses. Facilitation, antifacilitation, and rebound are still present and still depend on the transmitter inactivation rate. Thus, in the model, we can have a truly inexcitable membrane which nevertheless utilizes the active membrane conductances to provide facilitation or antifacilitation, and rebound. The simulated subthreshold pacemaker potentials are much more realistic with the increased capacitance, being lower in frequency and more natural in form.
In one case, the electronic model predicted behavior which was subsequently reported in real neurons. This was in respect to the interaction of synaptic potentials and pacemaker potential. It was noted in early experiments that when the model was set in a pacemaker mode, and periodic spikes were applied to the simulated inhibitory synapse, the pacemaker frequency could be modified; and, in fact, it would tend to lock on to the stimulus frequency. This produced a paradoxical effect whereby the frequency of spontaneous spikes was actually increased by increasing the frequency of inhibitory synaptic stimuli. At very low stimulus frequencies, the spontaneous pacemaker frequency was not appreciably perturbed. As the stimulus frequency was increased, and approached the basic pacemaker frequency, the latter tended to lock on and follow further increases in the stimulus frequency. When the stimulus frequency became too high for the pacemaker to follow, the latter decreased abruptly in frequency and locked on to the first subharmonic. As the stimulus frequency was further increased, the pacemaker frequency would increase, then skip to the next harmonic, then increase again, etc. This type of behavior was observed by Moore et al. [(23)] in Aplysia and reported at the San Diego Symposium for Biomedical Electronics shortly after it was observed by the author in the electronic model.
Thus, we have shown that an electronic analog with all parameters except membrane capacitance fixed at values close to those of Hodgkin and Huxley, can provide all of the normal threshold or axonal behavior and also all of the subthreshold somatic and dendritic behavior outlined on [page 7]. Whether or not this is of physiological significance, it certainly provides a unifying basis for construction of electronic neural analogs. Simple circuits, based on the Hodgkin-Huxley model and providing all of the aforementioned behavior, have been constructed with ten or fewer inexpensive transistors with a normal complement of associated circuitry [(18)]. In the near future we hope to utilize several models of this type to help assess the information-processing capabilities not only of individual neurons but also of small groups or networks of neurons.
REFERENCES
| 1. | Hagiwara, S., and Bullock, T. H. |
| “Intracellular Potentials in Pacemaker and Integrative Neurons of the Lobster Cardiac Ganglion,” | |
| J. Cell and Comp. Physiol. 50 (No. 1):25-48 (1957) | |
| 2. | Chalazonitis, N., and Arvanitaki, A., |
| “Slow Changes during and following Repetitive Synaptic Activation in Ganglion Nerve Cells,” | |
| Bull. Inst. Oceanogr. Monaco No. 1225:1-23 (1961) | |
| 3. | Hodgkin, A. L., Huxley, A. F., and Katz, B., |
| “Measurement of Current-Voltage Relations in the Membrane of the Giant Axon of Loligo,” | |
| J. Physiol. 116:424-448 (1952) | |
| 4. | Hagiwara, S., and Saito, N., |
| “Voltage-Current Relations in Nerve Cell Membrane of Onchidium verruculatum,” | |
| J. Physiol. 148:161-179 (1959) | |
| 5. | Hagiwara, S., and Saito, N., |
| “Membrane Potential Change and Membrane Current in Supramedullary Nerve Cell of Puffer,” | |
| J. Neurophysiol. 22:204-221 (1959) | |
| 6. | Hagiwara, S., |
| “Current-Voltage Relations of Nerve Cell Membrane,” | |
| “Electrical Activity of Single Cells,” | |
| Igakushoin, Hongo, Tokyo (1960) | |
| 7. | Bullock, T. H., |
| “Parameters of Integrative Action of the Nervous System at the Neuronal Level,” | |
| Experimental Cell Research Suppl. 5:323-337 (1958) | |
| 8. | Otani, T., and Bullock, T. H., |
| “Effects of Presetting the Membrane Potential of the Soma of Spontaneous and Integrating Ganglion Cells,” | |
| Physiological Zoology 32 (No. 2):104-114 (1959) | |
| 9. | Bullock, T. H., and Terzuolo, C. A., |
| “Diverse Forms of Activity in the Somata of Spontaneous and Integrating Ganglion Cells,” | |
| J. Physiol. 138:343-364 (1957) | |
| 10. | Bullock, T. H., |
| “Neuron Doctrine and Electrophysiology,” | |
| Science 129 (No. 3355):997-1002 (1959) | |
| 11. | Chalazonitis, N., and Arvanitaki, A., |
| “Slow Waves and Associated Spiking in Nerve Cells of Aplysia,” | |
| Bull. Inst. Oceanogr. Monaco No. 1224:1-15 (1961) | |
| 12. | Bullock, T. H., |
| “Properties of a Single Synapse in the Stellate Ganglion of Squid,” | |
| J. Neurophysiol. 11:343-364 (1948) | |
| 13. | Bullock, T. H., |
| “Neuronal Integrative Mechanisms,” | |
| “Recent Advances in Invertebrate Physiology,” | |
| Scheer, B. T., ed., Eugene, Oregon:Univ. Oregon Press 1957 | |
| 14. | Hodgkin, A. L., and Huxley, A. F., |
| “Currents Carried by Sodium and Potassium Ions through the Membrane of the Giant Axon of Loligo,” | |
| J. Physiol. 116:449-472 (1952) | |
| 15. | Hodgkin, A. L., and Huxley, A. F., |
| “The Components of Membrane Conductance in the Giant Axon of Loligo,” | |
| J. Physiol. 116:473-496 (1952) | |
| 16. | Hodgkin, A. L., and Huxley, A. F., |
| “The Dual Effect of Membrane Potential on Sodium Conductance in the Giant Axon of Loligo,” | |
| J. Physiol. 116:497-506 (1952) | |
| 17. | Hodgkin, A. L., and Huxley, A. F., |
| “A Quantitative Description of Membrane Current and its Application to Conduction and Excitation in Nerve,” | |
| J. Physiol. 117:500-544 (1952) | |
| 18. | Lewis, E. R., |
| “An Electronic Analog of the Neuron Based on the Dynamics of Potassium and Sodium Ion Fluxes,” | |
| “Neural Theory and Modeling,” | |
| R. F. Reiss, ed., Palo Alto, California:Stanford University Press, 1964 | |
| 19. | Eccles, J. C., |
| Physiology of Synapses, | |
| Berlin:Springer-Verlag, 1963 | |
| 20. | Grundfest, H., |
| “Excitation Triggers in Post-Junctional Cells,” | |
| “Physiological Triggers,” | |
| T. H. Bullock, ed., Washington, D.C.:American Physiological Society, 1955 | |
| 21. | Rall, W., |
| “Membrane Potential Transients and Membrane Time Constants of Motoneurons,” | |
| Exp. Neurol. 2:503-532 (1960) | |
| 22. | Araki, T., and Otani, T., |
| “The Response of Single Motoneurones to Direct Stimulation,” | |
| J. Neurophysiol. 18:472-485 (1955) | |
| 23. | Moore, G. P., Perkel, D. H., and Segundo, J. P., |
| “Stability Patterns in Interneuronal Pacemaker Regulation,” | |
| Proceedings of the San Diego Symposium for Biomedical Engineering, | |
| San Diego, California, 1963 | |
| 24. | Eccles, J. C., |
| The Neurophysiological Basis of Mind, | |
| Oxford:Clarendon Press, 1952 |
Fields and Waves in Excitable
Cellular Structures
R. M. STEWART
Space General Corporation
El Monte, California
“Study of living processes by the physiological method only proceeded laboriously behind the study of non-living systems. Knowledge about respiration, for instance, began to become well organized as the study of combustion proceeded, since this is an analogous operation....”
INTRODUCTION
The study of electrical fields in densely-packed cellular media is prompted primarily by a desire to understand more fully the details of brain mechanism and its relation to behavior. Our work has specifically been directed toward an attempt to model such structures and mechanisms, using relatively simple inorganic materials.
The prototype for such experiments is the “Lillie[1] iron-wire nerve model.” Over a hundred years ago, it had been observed that visible waves were produced on the surface of a piece of iron submerged in nitric acid when and where the iron is touched by a piece of zinc. After a short period of apparent fatigue, the wire recovers and can again support a wave when stimulated. Major support for the idea that such impulses are in fact directly related to peripheral nerve impulses came from Lillie around 1920. Along an entirely different line, various persons have noted the morphological and dynamic similarity of dendrites in brain and those which sometimes grow by electrodeposition of metals from solution. Gordon Pask [(17)], especially, has pointed to this similarity and has discussed in a general way the concomitant possibility of a physical model for the persistent memory trace.
By combining and extending such concepts and techniques, we hope to produce a macroscopic model of “gray matter,” the structural matrix of which will consist of a dense, homogeneously-mixed, conglomerate of small pellets, capable of supporting internal waves of excitation, of changing electrical behavior through internal fine-structure growth, and of forming temporal associations in response to peripheral shocks.
A few experimenters have subsequently pursued the iron-wire nerve-impulse analogy further, hoping thereby to illuminate the mechanisms of nerve excitation, impulse transmission and recovery, but interest has generally been quite low. It has remained fairly undisturbed in the text books and lecture demonstrations of medical students, as a picturesque aid to their formal education. On the outer fringes of biology, still less interest has been displayed; the philosophical vitalists would surely be revolted by the idea of such models of mind and memory, and at the other end of the scale, contemporary computer engineers generally assume that a nerve cell operates much too slowly to be of any value. This lack of interest is certainly due, in part, to success in developing techniques of monitoring individual nerve fibers directly to the point that it is just about as easy to work with large nerve fibers (and even peripheral and spinal junctions) as it is to work with iron wires. Under such circumstances, the model has only limited value, perhaps just to the extent that it emphasizes the role of factors other than specific molecular structure and local chemical reactions in the dynamics of nerve action.
When we leave the questions of impulse transmission on long fibers and peripheral junctions, however, and attempt to discuss the brain, there can be hardly any doubt that the development of a meaningful physical model technique would be of great value. Brain tissue is soft and sensitive, the cellular structures are small, tangled, and incredibly numerous. Therefore ([Young (24)]), “ ... physiologists hope that after having learned a lot about nerve-impulses in the nerves they will be able to go on to study how these impulses interact when they reach the brain. [But], we must not assume that we shall understand the brain only in the terms we have learned to use for the nerves. The function of nerves is to carry impulses—like telegraph wires. The functions of brains is something else.” But, confronted with such awesome experimental difficulties, with no comprehensive mathematical theory in sight, we are largely limited otherwise to verbal discourses, rationales and theorizing, a hopelessly clumsy tool for the development of an adequate understanding of brain function. A little over ten years ago [Sperry (19)] said, “Present day science is quite at a loss even to begin to describe the neural events involved in the simplest form of mental activity.” This situation has not changed much today. The development, study, and understanding of complex high-density cellular structures which incorporate characteristics of both the Lillie and Pask models may, it is hoped, alleviate this situation. There would also be fairly obvious technological applications for such techniques if highly developed and which, more than any other consideration, has prompted support for this work.
Experiments to date have been devised which demonstrate the following basic physical functional characteristics:
(1) Control of bulk resistivity of electrolytes containing closely-packed, poorly-conducting pellets
(2) Circulation of regenerative waves on closed loops
(3) Strong coupling between isolated excitable sites
(4) Logically-complete wave interactions, including facilitation and annihilation
(5) Dendrite growth by electrodeposition in “closed” excitable systems
(6) Subthreshold distributed field effects, especially in locally-refractory regions.
In addition, our attention has necessarily been directed to various problems of general experimental technique and choice of materials, especially as related to stability, fast recovery and long life. However, in order to understand the possible significance of, and motivation for such experiments, some related modern concepts of neurophysiology, histology and psychology will be reviewed very briefly. These concepts are, respectively:
- (1) Cellular structure in the central nervous system
- (2) Short-term or “ephemeral” memory
- (3) The synapse
- (4) Inhibition
- (5) Long-term memory traces or engram
- (6) Spatially-diffuse temporal association and learning.
SOME CONTEMPORARY CONCEPTS
Since we are attempting to duplicate processes other than chemical, per se, we will forego any reference to the extensive literature of neurochemistry. It should not be surprising though if, at the neglect of the fundamental biological processes of growth, reproduction and metabolism, it proves possible to imitate some learning mechanisms with grossly less complex molecular structures. There is also much talk of chemical versus electrical theories and mechanisms in neurophysiology. The distinction, when it can be made, seems to hinge on the question of the scale of size of significant interactions. Thus, “chemical” interactions presumably take place at molecular distances, possibly as a result of or subsequent to a certain amount of thermal diffusion. “Electrical” interactions, on the other hand, are generally understood to imply longer range or larger scale macroscopic fields.
1. Cellular Structure
The human brain contains approximately 10¹⁰ neurons to which the neuron theory assigns the primary role in central nervous activity. These cells occupy, however, a relatively small fraction of the total volume. There are, for example, approximately 10 times that number of neuroglia, cells of relatively indeterminate function. Each neuron (consisting of cell body, dendrites and, sometimes, an axon) comes into close contact with the dendrites of other neurones at some thousands of places, these synapses and “ephapses” being spaced approximately 5μ apart [(1)]. The total number of such apparent junctions is therefore of the order of 10¹³. In spite of infinite fine-structure variations when viewed with slightly blurred vision, the cellular structure of the brain is remarkably homogeneous. In the cortex, at least, the extensions of most cells are relatively short, and when the cortex is at rest, it appears from the large EEG alpha-rhythms that large numbers of cells beat together in unison. Quoting again from Sperry, “In short, current brain theory encourages us to try to correlate our subjective psychic experience with the activity of relatively homogeneous nerve cell units conducting essentially homogeneous impulses, through roughly homogeneous cerebral tissue.”
2. Short-Term Memory
A train of impulses simply travelling on a long fiber may, for example, be regarded as a short-term memory much in the same way as a delay line acts as a transient memory in a computer. A similar but slightly longer term memory may also be thought of to exist in the form of waves circulating in closed loops [(23)]. In fact, it is almost universally held today that most significant memory occurs in two basic interrelated ways. First of all, such a short-term circulating, reverberatory or regenerative memory which, however, could not conceivably persist through such things as coma, anesthesia, concussion, extreme cold, deep sleep and convulsive seizures and thus, secondly, a long-term memory trace which must somehow reside in a semipermanent fine-structural change. As [Hebb (9)] stated, “A reverbratory trace might cooperate with a structural change and carry the memory until the growth change is made.”
3. The Synapse
The current most highly regarded specific conception of the synapse is largely due to and has been best described by [Eccles (5)]: “ ... the synaptic connections between nerve cells are the only functional connections of any significance. These synapses are of two types, excitatory and inhibitory, the former type tending to make nerve cells discharge impulses, the other to suppress the discharge. There is now convincing evidence that in vertebrate synapses each type operates through specific chemical transmitter substances ...”. In response to a presentation by [Hebb (10)], Eccles was quoted as saying, “One final point, and that is if there is electrical interaction, and we have seen from Dr. Estable’s work the complexity of connections, and we now know from the electronmicroscopists that there is no free space, only 200 Å clefts, everywhere in the central nervous system, then everything should be electrically interacted with everything else. I think this is only electrical background noise and, that when we lift with specific chemical connections above that noise we get a significant operational system. I would say that there is electrical interaction but it is just a noise, a nuisance.” Eccles’ conclusions are primarily based on data obtained in the peripheral nervous system and the spinal cord. But there is overwhelming reason to expect that cellular interactions in the brain are an entirely different affair. For example, “The highest centres in the octopus, as in vertebrates and arthropods, contain many small neurons. This finding is such a commonplace, that we have perhaps failed in the past to make the fullest inquiry into its implications. Many of these small cells possess numerous processes, but no axon. It is difficult to see, therefore, that their function can be conductive in the ordinary sense. Most of our ideas about nervous functioning are based on the assumption that each neuron acts essentially as a link in some chain of conduction, but there is really no warrant for this in the case of cells with many short branches. Until we know more of the relations of these processes to each other in the neuropile it would be unwise to say more. It is possible that the effective part of the discharge of such cells is not as it is in conduction in long pathways, the internal circuit that returns through the same fiber, but the external circuit that enters other processes, ...” [(3)].
4. Inhibition
The inhibitory chemical transmitter substance postulated by Eccles has never been detected in spite of numerous efforts to do so. The mechanism(s) of inhibition is perhaps the key to the question of cellular interaction and, in one form or another, must be accounted for in any adequate theory.
Other rather specific forms of excitation and inhibition interaction have been proposed at one time or another. Perhaps the best example is the polar neuron of [Gesell (8)] and, more recently, [Retzlaff (18)]. In such a concept, excitatory and inhibitory couplings differ basically because of a macroscopic structural difference at the cellular level; that is, various arrangements or orientation of intimate cellular structures give rise to either excitation or inhibition.
5. Long-Term Memory
Most modern theories of semipermanent structural change (or engrams, as they are sometimes called) look either to the molecular level or to the cellular level. Various specific locales for the engram have been suggested, including [(1)] modifications of RNA molecular structure, [(2)] changes of cell size, synapse area or dendrite extensions, [(3)] neuropile modification, and [(4)] local changes in the cell membrane. There is, in fact, rather direct evidence of the growth of neurons or their dendrites with use and the diminution or atrophy of dendrites with disuse. The apical dendrite of pyramidal neurones becomes thicker and more twisted with continuing activity, nerve fibers swell when active, sprout additional branches (at least in the spinal cord) and presumably increase the size and number of their terminal knobs. As pointed out by [Konorski (11)], the morphological conception of plasticity according to which plastic changes would be related to the formation and multiplication of new synaptic junctions goes back at least as far as Ramon y Cajal in 1904. Whatever the substrate of the memory trace, it is, at least in adults, remarkably immune to extensive brain damage and as [Young (24)] has said: “ ... this question of the nature of the memory trace is one of the most obscure and disputed in the whole of biology.”
6. Field Effects and Learning
First, from [Boycott and Young (3)], “The current conception, on which most discussions of learning still concentrate, is that the nervous system consists essentially of an aggregate of chains of conductors, linked at key points by synapses. This reflex conception, springing probably from Cartesian theory and method, has no doubt proved of outstanding value in helping us to analyse the actions of the spinal cord, but it can be argued that it has actually obstructed the development of understanding of cerebral function.”
Most observable evidence of learning and memory is extremely complex and its interpretation full of traps. Learning in its broadest sense might be detected as a semipermanent change of behavior pattern brought about as a result of experience. Within that kind of definition, we can surely identify several distinctly different types of learning, presumably with distinctly different kinds of mechanisms associated with each one. But, if we are to stick by our definition of a condition of semipermanent change of behavior as a criterion for learning, then we may also be misled into considering the development of a neurosis, for example, as learning, or even a deep coma as learning.
When we come to consider field effects, current theories tend to get fairly obscure, but there seems to be an almost universal recognition of the fact that such fields are significant. For example, [Morrell (16)] says in his review of electrophysiological contributions to the neural basis of learning, “A growing body of knowledge (see reviews by Purpura, Grundfest, and Bishop) suggests that the most significant integrative work of the central nervous system is carried on in graded response elements—elements in which the degree of reaction depends upon stimulus intensity and is not all-or-none, which have no refractory period and in which continuously varying potential changes of either sign occur and mix and algebraically sum.” [Gerard (7)] also makes a number of general comments along these lines. “These attributes of a given cell are, in turn, normally controlled by impulses arising from other regions, by fields surrounding them—both electric and chemical—electric and chemical fields can strongly influence the interaction of neurones. This has been amply expounded in the case of the electric fields.”
Learning situations involving “punishment” and “reward” or, subjectively, “pain” and “pleasure” may very likely be associated with transient but structurally widespread field effects. States of distress and of success seem to exert a lasting influence on behavior only in relation to simultaneous sensory events or, better yet, sensory events just immediately preceding in time. For example, the “anticipatory” nature of a conditioned reflex has been widely noted [(21)]. From a structural point of view, it is as if recently active sites regardless of location or function were especially sensitive to extensive fields. There is a known inherent electrical property of both nerve membrane and passive iron surface that could hold the answer to this mechanism of spatially-diffuse temporal association; namely, the surface resistance drops to less than 1 per cent of its resting value during the refractory period which immediately follows activation.
EXPERIMENTAL TECHNIQUE
In almost all experiments, the basic signal-energy mechanism employed has been essentially that one studied most extensively by [Lillie (12)], [Bonhoeffer (2)], [Yamagiwa (22)], [Matumoto and Goto (14)] and others, i.e., activation, impulse propagation and recovery on the normally passive surface of a piece of iron immersed in nitric acid or of cobalt in chromic acid [(20)]. The iron we have used most frequently is of about 99.99% purity, which gives performance more consistent than but similar to that obtained using cleaned “coat-hanger” wires. The acid used most frequently by us is about 53-55% aqueous solution by weight, substantially more dilute than that predominantly used by previous investigators. The most frequently reported concentration has been 68-70%, a solution which is quite stable and, hence, much easier to work with in open containers than the weaker solutions, results in very fast waves but gives, at room temperatures, a very long refractory period (typically, 15 minutes). A noble metal (such as silver, gold or platinum) placed in contact with the surface of the iron has a stabilizing effect [(14)] presumably through the action of local currents and provides a simple and useful technique whereby, with dilution, both stability and fast recovery (1 second) can be achieved in simple demonstrations and experiments.
Experiments involving the growth by electrodeposition and study of metallic dendrites are done with an eye toward electrical, physical and chemical compatibility with the energy-producing system outlined above. Best results to date (from the standpoints of stability, non-reactivity, and morphological similarity to neurological structures) have been obtained by dissolving various amounts of gold chloride salt in 53-55% HNO₃.
An apparatus has been devised and assembled for the purpose of containing and controlling our primary experiments. ([See Figure 1]). Its two major components are a test chamber (on the left in [Figure 1]) and a fluid exchanger (on the right). In normal operation the test chamber, which is very rigid and well sealed after placing the experimental assembly inside, is completely filled with electrolyte (or, initially, an inert fluid) to the exclusion of all air pockets and bubbles. Thus encapsulated, it is possible to perform experiments which would otherwise be impossible due to instability. The instability which plagues such experiments is manifested in copious generation of bubbles on and subsequent rapid disintegration of all “excitable” material (i.e., iron). Preliminary experiments indicated that such “bubble instability” could be suppressed by constraining the volume available to expansion. In particular, response and recovery times can now be decreased substantially and work can proceed with complex systems of interest such as aggregates containing many small iron pellets.
The test chamber is provided with a heater (and thermostatic control) which makes possible electrochemical impulse response and recovery times comparable to those of the nervous system (1 to 10 msec). The fluid-exchanger is so arranged that fluid in the test chamber can be arbitrarily changed or renewed by exchange within a rigid, sealed, completely liquid-filled (“isochoric”) loop. Thus, stability can be maintained for long periods of time and over a wide variety of investigative or operating conditions.
Most of the parts of this apparatus are made of stainless steel and are sealed with polyethylene and teflon. There is a small quartz observation window on the test chamber, two small lighting ports, a pressure transducer, thermocouple, screw-and-piston pressure actuator and umbilical connector for experimental electrical inputs and outputs.
BASIC EXPERIMENTS
The basic types of experiments described in the following sections are numbered for comparison to correspond roughly to related neurophysiological concepts summarized in the previous section.
1. Cellular Structure
The primary object of our research is the control and determination of dynamic behavior in response to electrical stimulation in close-packed aggregates of small pellets submerged in electrolyte. Typically, the aggregate contains (among other things) iron and the electrolyte contains nitric acid, this combination making possible the propagation of electrochemical surface waves of excitation through the body of the aggregate similar to those of the Lillie iron-wire nerve model. The iron pellets are imbedded in and supported by a matrix of small dielectric (such as glass) pellets. Furthermore, with the addition of soluble salts of various noble metals to the electrolyte, long interstitial dendritic or fibrous structures of the second metal can be formed whose length and distribution change by electrodeposition in response to either internal or externally generated fields.
Figure 1—Test chamber and
fluid exchanger
Coupling between isolated excitable (iron) sites is greatly affected by the fine structure and effective bulk resistivity of the glass and fluid medium which supports and fills the space between such sites. In general ([see Section 3, following]) it is necessary, to promote strong coupling between small structures, to impede the “short-circuit” return flow of current from an active or excited surface, through the electrolyte and back through the dendritic structure attached to the same excitable site. This calls for control (increase) of the bulk resistivity, preferably by means specifically independent of electrolyte composition, which relates to and affects surface phenomena such as recovery (i.e., the “refractory” period). [Figure 2] illustrates the way in which this is being done, i.e., by appropriate choice of particle size distributions. The case illustrated shows the approximate proper volume ratios for maximum resistivity in a two-size-phase random mixture of spheres.
2. Regenerative Loops
[Figure 3] shows an iron loop (about 2-inch diameter) wrapped with a silver wire helix which is quite stable in 53-55% acid and which will easily support a circulating pattern of three impulses. For demonstration, unilateral waves can be generated by first touching the iron with a piece of zinc (which produces two oppositely travelling waves) and then blocking one of them with a piece of platinum or a small platinum screen attached to the end of a stick or wand. Carbon blocks may also be used for this purpose.
The smallest regenerative or reverberatory loop which we are at present able to devise is about 1 mm in diameter. Multiple waves, as expected, produce stable patterns in which all impulses are equally spaced. This phenomenon can be related to the slightly slower speed characteristic of the relative refractory period as compared with a more fully recovered zone.
Figure 2—Conductivity control—mixed pellet-size aggregates
Figure 3—Regenerative or reverberatory loop
3. Strong Coupling
If two touching pieces of iron are placed in a bath of nitric acid, a wave generated on one will ordinarily spread to the other. As is to be expected, a similar result is obtained if the two pieces are connected through an external conducting wire. However, if they are isolated, strong coupling does not ordinarily occur, especially if the elements are small in comparison with a “critical size,” σ/ρ where σ is the surface resistivity of passive iron surface (in Ω-cm²) and ρ is the volume resistivity of the acid (in Ω-cm). A simple and informative structure which demonstrates the essential conditions for strong electrical coupling between isolated elements of very small size may be constructed as shown in [Figure 4]. The dielectric barrier insures that charge transfer through one dipole must be accompanied by an equal and opposite transfer through the surfaces of the other dipole. If the “inexcitable” silver tails have sufficiently high conductance (i.e., sufficiently large surface area, hence preferably, dendrites), strong coupling will occur, just as though the cores of the two pieces of iron were connected with a solid conducting wire.
Figure 4
Figure 5—Electrochemical excitatory-inhibitory
interaction cell
4. Inhibitory Coupling
If a third “dipole” is inserted through the dielectric membrane in the opposite direction, then excitation of this isolated element tends to inhibit the response which would otherwise be elicited by excitation of one of the parallel dipoles. [Figure 5] shows the first such “logically-complete” interaction cell successfully constructed and demonstrated. It may be said to behave as an elementary McCulloch-Pitts neuron [(15)]. Further analysis shows that similar structures incorporating many dipoles (both excitatory and inhibitory) can be made to behave as general “linear decision functions” in which all input weights are approximately proportional to the total size or length of their corresponding attached dendritic structures.
5. Dendrite Growth
[Figure 6] shows a sample gold dendrite grown by electrodeposition (actual size, about 1 mm) from a 54% nitric acid solution to which gold chloride was added. When such a dendrite is attached to a piece of iron (both submerged), activation of the excitable element produces a field in such a direction as to promote further growth of the dendritic structure. Thus, if gold chloride is added to the solution used in the elementary interaction cells described above, all input influence “weights” tend to increase with use and, hence, produce a plasticity of function.
6. Field Effects in Locally-Refractory Regions
Our measurements indicate that, during the refractory period following excitation, the surface resistance of iron in nitric acid drops to substantially less than 1% of its resting value in a manner reminiscent of nerve membranes [(4)]. Thus, if a distributed or gross field exists at any time throughout a complex cellular aggregate, concomitant current densities in locally-refractive regions will be substantially higher than elsewhere and, if conditions appropriate to dendrite growth exist (as described above) growth rates in such regions will also be substantially higher than elsewhere. It would appear that, as a result, recently active functional couplings (in contrast to those not associated with recent neural activity) should be significantly altered by widely distributed fields or massive peripheral shocks. This mechanism might thus explain the apparent ability of the brain to form specific temporal associations in response to spatially-diffuse effects such as are generated, for example, by the pain receptors.
(a)
(b)
Figure 6—Dendritic structures, living and non-living. (a) Cat dendrite trees (from Bok, “Histonomy of the Cerebral Cortex,” Elsevier, 1959); (b) Electrodeposited gold dendrite tree.
SUMMARY
An attempt is being made to develop meaningful electrochemical model techniques which may contribute toward a clearer understanding of cortical function. Two basic phenomena are simultaneously employed which are variants of (1) the Lillie iron-wire nerve model, and (2) growth of metallic dendrites by electrodeposition. These phenomena are being induced particularly within dense cellular aggregates of various materials whose interstitial spaces are flooded with liquid electrolyte.
REFERENCES
| 1. | Bok, S. T., |
| “Histonomy of the Cerebral Cortex,” | |
| Amsterdam, London:Elsevier Publishing Co., New York:Princeton, 1959 | |
| 2. | Bonhoeffer, K. F., |
| “Activation of Passive Iron as a Model for the Excitation of Nerve,” | |
| J. Gen. Physiol. 32:69-91 (1948). | |
| This paper summarizes work carried out during 1941-1946 at the University of Leipzig, and published during the war years in German periodicals. | |
| 3. | Boycott, B. B., and Young, J. Z., |
| “The Comparative Study of Learning,” | |
| S. E. B. Symposia, No. IV | |
| “Physiological Mechanisms in Animal Behavior,” | |
| Cambridge: University Press, USA:Academic Press, Inc., 1950 | |
| 4. | Cole, K. S., and Curtis, H. J., |
| “Electric Impedance of the Squid Giant Axon During Activity,” | |
| J. Gen. Physiol. 22:649-670 (1939) | |
| 5. | Eccles, J. C., |
| “The Effects of Use and Disuse of Synaptic Function,” | |
| “Brain Mechanisms and Learning—A Symposium,” | |
| organized by the Council for International Organizations of Medical Science, Oxford:Blackwell Scientific Publications, 1961 | |
| 6. | Franck, U. F., |
| “Models for Biological Excitation Processes,” | |
| “Progress in Biophysics and Biophysical Chemistry,” | |
| J. A. V. Butler, ed., London and New York:Pergamon Press, pp. 171-206, 1956 | |
| 7. | Gerard, R. W., |
| “Biological Roots of Psychiatry,” | |
| Science 122 (No. 3162):225-230 (1955) | |
| 8. | Gesell, R., |
| “A Neurophysiological Interpretation of the Respiratory Act,” | |
| Ergedn. Physiol. 43:477-639 (1940) | |
| 9. | Hebb, D. O., |
| “The Organization of Behavior, A Neuropsychological Theory,” | |
| New York:John Wiley and Sons, 1949 | |
| 10. | Hebb, D. O., |
| “Distinctive Features of Learning in the Higher Animal,” | |
| “Brain Mechanisms and Learning—A Symposium,” | |
| organized by the Council for International Organizations of Medical Science, Oxford:Blackwell Scientific Publications, 1961 | |
| 11. | Konorski, J., |
| “Conditioned Reflexes and Neuron Organization,” | |
| Cambridge:Cambridge University Press, 1948 | |
| 12. | Lillie, R. S., |
| “Factors Affecting the Transmission and Recovery in the Passive Iron Nerve Model,” | |
| J. Gen. Physiol. 4:473 (1925) | |
| 13. | Lillie, R. S., |
| Biol. Rev. 16:216 (1936) | |
| 14. | Matumoto, M., and Goto, K., |
| “A New Type of Nerve Conduction Model,” | |
| The Gurma Journal of Medical Sciences 4(No. 1) (1955) | |
| 15. | McCulloch, W. S., and Pitts, W., |
| “A Logical Calculus of the Ideas Immanent in Nervous Activity,” | |
| Bulletin of Mathematical Biophysics 5:115-133 (1943) | |
| 16. | Morrell, F., |
| “Electrophysiological Contributions to the Neural Basis of Learning,” | |
| Physiological Reviews 41(No. 3) (1961) | |
| 17. | Pask, G., |
| “The Growth Process Inside the Cybernetic Machine,” | |
| Proc. 2nd Congress International Association Cybernetics, Gauthier-Villars, Paris:Namur, 1958 | |
| 18. | Retzlaff, E., |
| “Neurohistological Basis for the Functioning of Paired Half-Centers,” | |
| J. Comp. Neurology 101:407-443 (1954) | |
| 19. | Sperry, R. W., |
| “Neurology and the Mind-Brain Problem,” | |
| Amer. Scientist 40(No. 2): 291-312 (1952) | |
| 20. | Tasaki, I., and Bak, A. F., |
| J. Gen. Physiol. 42:899 (1959) | |
| 21. | Thorpe, W. H., |
| “The Concepts of Learning and Their Relation to Those of Instinct,” | |
| S. E. B. Symposia, No. IV, | |
| “Physiological Mechanisms in Animal Behavior,” | |
| Cambridge:University Press, USA:Academic Press, Inc., 1950 | |
| 22. | Yamagiwa, K., |
| “The Interaction in Various Manifestations (Observations on Lillie’s Nerve Model),” | |
| Jap. J. Physiol. 1:40-54 (1950) | |
| 23. | Young, J. Z., |
| “The Evolution of the Nervous System and of the Relationship of Organism and Environment,” | |
| G. R. de Beer, ed., | |
| “Evolution,” | |
| Oxford:Clarendon Press, pp. 179-204, 1938 | |
| 24. | Young, J. Z., |
| “Doubt and Certainty in Science, A Biologist’s Reflections on the Brain,” | |
| New York:Oxford Press, 1951 |
Multi-Layer Learning Networks
R. A. STAFFORD
Philco Corp., Aeronutronic Division
Newport Beach, California
INTRODUCTION
This paper is concerned with the problem of designing a network of linear threshold elements capable of efficiently adapting its various sets of weights so as to produce a prescribed input-output relation. It is to accomplish this adaptation by being repetitively presented with the various inputs along with the corresponding desired outputs. We will not be concerned here with the further requirement of various kinds of ability to “generalize”—i.e., to tend to give correct outputs for inputs that have not previously occurred when they are similar in some transformed sense to other inputs that have occurred.
In putting forth a model for such an adapting or “learning” network, a requirement is laid down that the complexity of the adaption process in terms of interconnections among elements needed for producing appropriate weight changes, should not greatly exceed that already required to produce outputs from inputs with a static set of weights. In fact, it has been found possible to use the output-from-input computing capacity of the network to help choose proper weight changes by observing the effect on the output of a variety of possible weight changes.
No attempt is made here to defend the proposed network model on theoretical grounds since no effective theory is known at present. Instead, the plausibility of the various aspects of the network model, combined with empirical results must suffice.
SINGLE ELEMENTS
To simplify the problem it is assumed that the network receives a set of two-valued inputs, x₁, x₂, ..., xₙ, and is required to produce only a single two-valued output, y. It is convenient to assign the numerical quantities +1 and -1 to the two values of each variable.
The simplest network would consist of a single linear threshold element with a set of weights, c₀, c₁, c₂, ..., cₙ. These determine the output-input relation or function so that y is +1 or -1 according as the quantity, c₀ + c₁x₁ + c₂x₂ + ... + cₙxₙ, is positive or not, respectively. It is possible for such a single element to exhibit an adaptive behavior as follows. If, for a given set, x₁, x₂, ..., xₙ, the output, y, is correct, then make no changes to the weights. Otherwise change the weights according to the equations
- Δc₀ = y*
- Δcᵢ = y*xᵢ, i = 1,2, ...,n
- where y* is the desired output.
It has been shown by a number of people that the weights of such an element are assured of arriving at a set of values which produce the correct output-input relation after a sufficient number of errors, provided that such a set exists. An upper bound on the number of possible errors can be given which depends only on the initial weight values and the logical function to be learned. This does not, however, solve our network problem for two reasons.
First, as the number, n, of inputs gets large, the number of errors to be expected for most functions which can be learned increases to unreasonable values. For example, for n = 6, most such functions result in 500 to 1000 errors compared to an average of 32 errors to be expected in a perfect learning device.
Second, and more important, the fraction of those logical functions which can be generated in a single element becomes vanishingly small as n increases. For example, at n = 6 less than one in each three trillion logical functions is so obtainable.
NETWORKS OF ELEMENTS
It can be demonstrated that if a sufficiently large number of linear threshold elements is used, with the outputs of some being the inputs of others, then a final output can be produced which is any desired logical function of the inputs. The difficulty in such a network lies in the fact that we are no longer provided with a knowledge of the correct output for each element, but only for the final output. If the final output is incorrect there is no obvious way to determine which sets of weights should be altered.
As a result of considerable study and experimentation at Aeronutronic, a network model has been evolved which, it is felt, will get around these difficulties. It consists of four basic features which will now be described.
Positive Interconnecting Weights
It is proposed that all weights in elements attached to inputs which come from other elements in the network be restricted to positive values. (Weights attached to the original inputs to the network, of course, must be allowed to be of either sign.) The reason for such a restriction is this. If element 1 is an input to element 2 with weight c₁₂, element 2 to element 3 with weight c₂₃, etc., then the sign of the product, c₁₂c₂₃ ..., gives the sense of the effect of a change in the output of element 1 on the final element in the chain (assuming this is the only such chain between the two elements). If these various weights were of either possible sign, then a decision as to whether or not to change the output in element 1 to help correct an error in the final element would involve all weights in the chain. Moreover, since there would in general be a multiplicity of such chains, the decision is rendered impossibly difficult.
The above restriction removes this difficulty. If the output of any element in the network is changed, say, from -1 to +1, the effect on the final element, if it is affected at all, is in the same direction.
It should be noted that this restriction does not seriously affect the logical capabilities of a network. In fact, if a certain logical function can be achieved in a network with the use of weights of unrestricted sign, then the same function can be generated in another network with only positive interconnecting weights and, at worst, twice the number of elements. In the worst case this is done by generating in the restricted network both the output and its complement for each element of the unrestricted network. (It is assumed that there are no loops in the network.)
A Variable Bias
The central problem in network learning is that of determining, for a given input, the set of elements whose outputs can be altered so as to correct the final element, and which will do the least amount of damage to previous adaptations to other inputs. Once this set has been determined, the incrementing rule given for a single element will apply in this case as well (subject to the restriction of leaving interconnecting weights positive), since the desired final output coincides with that desired for each of the elements to be changed (because of positive interconnecting weights).
In the process of arriving at such a decision three factors need to be considered. Elements selected for change should tend to be those whose output would thereby be affected for a minimum number of other possible inputs. At the same time it should be ascertained that a change in each of the elements in question does indeed contribute significantly towards correcting the final output. Finally, a minimum number of such elements should be used.
It would appear at first that this kind of decision is impossible to achieve if the complexity of the decision apparatus is kept comparable to that of the basic input-output network as mentioned earlier. However, in the method to be described it is felt that a reasonable approximation to these requirements will be achieved without an undue increase in complexity.
It is assumed that in addition to its normal inputs, each element receives a variable input bias which we can call b. The output of every element should then be determined by the sign of the usual weighted sum of its inputs plus this bias quantity. This bias is to be the same for each element of the network. If b = 0 the network will behave as before. However, if b is increased gradually, various elements throughout the network will commence changing from -1 to +1, with one or a few changing at any one time as a rule. If b is decreased, the opposite will occur.
Now suppose that for a given input the final output ought to be +1 but actually is -1. Assume that b is then raised so high that this final output is corrected. Then commence a gradual decline in b. Various elements may revert to -1, but until the final output does, no weights are changed. When the final output does revert to -1, it is due to an element’s having a sum (weighted sum plus bias) which just passed down through zero. This then caused a chain effect of changing elements up to the final element, but presumably this element is the only one possessing a zero sum. This can then be the signal for the weights on an element to change—a change of final output from right to wrong accompanied simultaneously by a zero sum in the element itself.
After such a weight change, the final output will be correct once more and the bias can again proceed to fall. Before it reaches zero, this process may occur a number of times throughout the network. When the bias finally stands at zero with the final output correct, the network is ready for the next input. Of course if -1 is desired, the bias will change in the opposite direction.
It is possible that extending the weight change process a little past the zero bias level may have beneficial results. This might increase the life expectancy of each learned input-output combination and thereby reduce the total number of errors. This is because the method used above can stop the weight correction process so that even though the final output is correct, some elements whose output are essential to the final output have sums close to zero, which are easily changed by subsequent weight changes.
It will be noted that this method conforms to all three considerations mentioned previously. First, by furnishing each element the same bias, and by not changing weights until the final output becomes incorrect with dropping bias, there is a strong tendency to select elements which, with b = 0, would have sums close to zero. But the size of the sum in an element is a good measure of the amount of damage done to an element for other inputs if its current output is to be changed. Second, it is obvious that each element changed has had a demonstrable effect on the final output. Finally, there will be a clear tendency to change only a minimum of elements because changes never occur until the output clearly requires a change.
On the other hand this method requires little more added complexity to the network than it already has. Each element requires a bias, an error signal, and the desired final output, these things being uniform for all elements in a network. Some external device must manipulate the bias properly, but this is a simple behavior depending only on an error signal and the desired final output—not on the state of individual elements in the network. What one has, then, is a network consisting of elements which are nearly autonomous as regards their decisions to change weights. Such a scheme appears to be the only way to avoid constructing a central weight-change decision apparatus of great complexity. This rather sophisticated decision is made possible by utilizing the computational capabilities the network already possesses in producing outputs from inputs.
It should be noted here that this varying bias method requires that the variable bias be furnished to just those elements which have variable weights and to no others. Any fixed portion of the network, such as preliminary layers or final majority function for example, must operate independently of the variable bias. Otherwise, the final output may go from right to wrong as the bias moves towards zero and no variable-weight element be to blame. In such a case the network would be hung up.
Logical Redundancy in the Network
A third aspect of the network model is that for all the care taken in the previous steps, they will not suffice in settling quickly to a set of weights that will generate the required logical function unless there is a great multiplicity of ways in which this can be done. This is to say that a learning network needs to have an excess margin of weights and elements beyond the minimum required to generate the functions which are to be learned.
This is analogous to the situation that prevails for a single element as regards the allowed range of values on its weights. It can be shown for example, that any function for n=6 that can be generated by a single element can be obtained with each weight restricted to the range of integer values -9,-8, ..., +9. Yet no modification of the stated weight change rule is known which restricts weight values to these and yet has any chance of ever being learned for most functions.
Fatigued Elements
It would appear from some of the preliminary results of network simulations that it may be useful to have elements become “fatigued” after undergoing an excessive number of weight changes. Experiments have been performed on simplifications of the model described so far which had the occasional result that a small number of elements came to a state where they received most of the weight increments, much to the detriment of the learning process. In such cases the network behaves as if it were composed of many fewer adjustable elements. In a sense this is asking each element to maintain a record of the data it is being asked to store so that it does not attempt to exceed its own information capacity.
It is not certain just how this fatigue factor should enter in the element’s actions, but if it is to be compatible with the variable bias method, this fatigue factor must enter into the element’s response to a changing bias. Once an element changes state with zero sum at the same time that the final output becomes wrong, incrementing must occur if the method is to work. Hence a “fatigued” element must respond less energetically to a change of bias, perhaps with a kind of variable factor to be multiplied by the bias term.
NETWORK STRUCTURE
It is felt that the problem of selecting the structure of interconnections for a network is intimately connected to the previously mentioned problem of generalization. Presumably a given type of generalization can be obtained by providing appropriate fixed portions of the network and an appropriate interconnection structure for the variable portion. However, for very large networks, it is undoubtedly necessary to restrict the complexity so that it can be specified by relatively simple rules. Since very little is known about this quite important problem, no further discussion will be attempted here.
COMPUTER SIMULATION RESULTS
A computer simulation of some of the network features previously described has been made on an IBM 7090. Networks with an excess of elements and with only positive interconnecting weights were used. However, in place of the variable bias method, a simple choice of the element of sum closest to, and on the wrong side of, zero was made without regard to the effectiveness of the element in correcting the final output. No fatigue factors were used.
The results of these simulations are very encouraging, but at the same time indicate the need for the more sophisticated methods. No attempt will be made here to describe the results completely.
In one series of learning experiments, a 22-element network was used which had three layers, 10 elements on the first, 11 on the second, and 1 on the third. The single element on the third was the final output, and was a fixed majority function of the 11 elements in the second layer. These in turn each received inputs from each of the 10 on the first layer and from each of the 6 basic inputs. The 10 on the first layer each received only the 6 basic inputs. A set of four logical functions, A, B, C, and D, was used. Function A was actually a linear threshold function which could be generated by the weights 8, 7, 6, 5, 4, 3, 2, functions B and C were chosen by randomly filling in a truth table, while D was the parity function.
| A | B | C | D | ||||
|---|---|---|---|---|---|---|---|
| r | e | r | e | r | e | r | e |
| 5 | 54 | 8 | 100 | 11 | 101 | 4 | 52 |
| 4 | 37 | 9 | 85 | 4 | 60 | 5 | 62 |
| 4 | 44 | 6 | 72 | 9 | 85 | 6 | 56 |
Table I gives the results of one series of runs with these functions and this network, starting with various random initial weights. The quantity, r, is the number of complete passes through the 64-entry truth table before the function was completely learned, while e is the total number of errors made. In evaluating the results it should be noted that an ideal learning device would make an average of 32 errors altogether on each run. The totals recorded in these runs are agreeably close to this ideal. As expected, the linear threshold function is the easiest to learn, but it is surprising that the parity function was substantially easier than the two randomly chosen functions. [Table II] gives a chastening result of the same experiment with all interconnecting weights removed except that the final element is a fixed majority function of the other 21 elements. Thus there was adaptation on one layer only. As can be seen [Table I] is hardly better than [Table II] so that the value of variable interconnecting weights was not being fully realized. In a later experiment the number of elements was reduced to 12 elements and the same functions used. In this case the presence of extra interconnecting weights actually proved to be a hindrance! However a close examination of the incrementing process brought out the fact that the troublesome behavior was due to the greater chance of having only a few (often only one) elements do nearly all the incrementing. It is expected that the use of the additional refinements discussed herein will produce a considerable improvement in bringing out the full power of adaptation in multiple layers of a network.
| A | B | C | D | ||||
|---|---|---|---|---|---|---|---|
| r | e | r | e | r | e | r | e |
| 7 | 47 | 18 | 192 | 8 | 110 | 4 | 48 |
| 3 | 40 | 7 | 69 | 10 | 98 | 6 | 68 |
| 4 | 43 | 7 | 82 | 4 | 47 | 6 | 46 |
FUTURE PROBLEMS
Aside from the previous question of deciding on network structure, there are several other questions that remain to be studied in learning networks.
There is the question of requiring more than a single output from a network. If, say, two outputs are required for a given input, one +1 and the other -1, this runs into conflict with the incrementing process. Changes that aid one output may act against the other. Apparently the searching process depicted before with a varying bias must be considerably refined to find weight changes which act on all the outputs in the required way. This is far from an academic question because there will undoubtedly be numerous cases in which the greatest part of the input-output computation will have shared features for all output variables. Only at later levels do they need to be differentiated. Hence it is necessary to envision a single network producing multiple outputs rather than a separate network for each output variable if full efficiency is to be achieved.
Another related question is that of using input variables that are either many-, or continuous-, valued rather than two-valued. No fundamental difficulties are discernible in this case, but the matter deserves some considerable study and experimentation.
Another important question involves the use of a succession of inputs for producing an output. That is, it may be useful to allow time to enter into the network’s logical action, thus giving it a “dynamic” as well as “static” capability.
Adaptive Detection of Unknown
Binary Waveforms
J. J. Spilker, Jr.
Philco Western Development Laboratories
Palo Alto, California
This work was supported by the Philco WDL Independent Development Program. This paper, submitted after the Symposium, represents a more detailed presentation of some of the issues raised in the discussion sessions at the Symposium and hence, constitutes a worthwhile addition to the Proceedings.
INTRODUCTION
One of the most important objectives in processing a stream of data is to determine and detect the presence of any invariant or quasi-invariant “features” in that data stream. These features are often initially unknown and must be “learned” from the observations. One of the simplest features of this form is a finite length signal which occurs repetitively, but not necessarily periodically with time, and has a waveshape that remains invariant or varies only slowly with time.
In this discussion, we assume that the data stream has been pre-processed, perhaps by a detector or discriminator, so as to exhibit this type of repetitive (but unknown) waveshape or signal structure. The observed signal, however, is perturbed by additive noise or other disturbances. It is desired to separate the quasi-invariance of the data from the truly random environment. The repetitive waveform may represent, for example, the transmission of an unknown sonar or radar, a pulse-position modulated noise-like waveform, or a repeated code word.
The problem of concern is to estimate the signal waveshape and to determine the time of each signal occurrence. We limit this discussion to the situation where only a single repetitive waveform is present and the signal sample values are binary. The observed waveform is assumed to be received at low signal-to-noise ratio so that a single observation of the signal (even if one knew precisely the arrival time) is not sufficient to provide a good estimate of the signal waveshape. The occurrence time of each signal is assumed to be random.
THE ADAPTIVE DETECTION MACHINE
The purpose of this note is to describe very briefly a machine[2] which has been implemented to recover the noise-perturbed binary waveform. A simplified block diagram of the machine is shown in [Figure 1]. The experimental machine has been designed to operate on signals of 10³ samples duration.
Each analog input sample enters the machine at left and may either contain a signal sample plus noise or noise alone. In order to permit digital operation in the machine, the samples are quantized in a symmetrical three-level quantizer. The samples are then converted to vector form, e.g., the previous 10³ samples form the vector components. A new input vector, Y⁽ⁱ⁾, is formed at each sample instant.
Define the signal sample values as s₁, s₂, ..., sₙ. The observed vector Y⁽ⁱ⁾ is then either (a) perfectly centered signal plus noise, (b) shifted signal plus noise, or (c) noise alone.
![]() | (s₁, s₂, ..., sₙ) + (n₁, n₂, ..., nₙ) | (a) | |
| (Y⁽ⁱ⁾)ᵗ = | (0, ..., s₁, s₂, ..., sₙ₋ⱼ) + (n₁, n₂, ..., nₙ) | (b) | |
| (0 ... 0) + (n₁, n₂, ..., nₙ) | (c) |
At each sample instant, two measurements are made on the input vector, an energy measurement ‖Y⁽ⁱ⁾‖² and a polarity coincidence cross-correlation with the present estimate of the signal vector stored in memory. If the weighted sum of the energy and cross-correlation measurements exceeds the present threshold value Γᵢ, the input vector is accepted as containing the signal (properly shifted in time), and the input vector is added to the memory. The adaptive memory has 2Q levels, 2Q-1 positive levels, 1 zero level and 2Q-1-1 negative levels. New contributions are made to the memory by normal vector addition except that saturation occurs when a component value is at the maximum or minimum level.
The acceptance or rejection of a given input vector is based on a hypersphere decision boundary. The input vector is accepted if the weighted sum γᵢ exceeds the threshold Γᵢ
γᵢ = Y⁽ⁱ⁾∙M⁽ⁱ⁾ + α‖Y⁽ⁱ⁾‖² ⩾ Γᵢ.
Figure 1—Block diagram of the adaptive binary waveform detector
Geometrically, we see that the input vector is accepted if it falls on or outside of a hypersphere centered at
| - M⁽ⁱ⁾ | |
| C⁽ⁱ⁾ = | —— |
| 2α |
having radius squared
| Γ⁽ⁱ⁾ | ‖M⁽ⁱ⁾‖² | ||
| [r⁽ⁱ⁾]² = | —— | + | ———— . |
| α | (2α)² |
Both the center and radius of this hypersphere change as the machine adapts. The performance and optimality of hypersphere-type decision boundaries have been discussed in related work by Glaser[3] and Cooper.[4]
The threshold value, Γᵢ, is adapted so that it increases if the memory becomes a better replica of the signal with the result that γᵢ increases. On the other hand, if the memory is a poor replica of the signal (for example, if it contains noise alone), it is necessary that the threshold decay with time to the point where additional acceptances can modify the memory structure.
The experimental machine is entirely digital in operation and, as stated above, is capable of recovering waveforms of up to 10³ samples in duration. In a typical experiment, one might attempt to recover an unknown noise-perturbed, pseudo-random waveform of up to 10³ bits duration which occurs at random intervals. If no information is available as to the signal waveshape, the adaptive memory is blank at the start of the experiment.
In order to illustrate the operation of the machine most clearly, let us consider a repetitive binary waveform which is composed of 10³ bits of alternate “zeros” and “ones.” A portion of this waveform is shown in [Figure 2a]. The waveform actually observed is a noise-perturbed version of this waveform shown in [Figure 2b] at-6 db signal-to-noise ratio. The exact sign of each of the signal bits obviously could not be accurately determined by direct observation of [Figure 2b].
(a) Binary signal
(b) Binary signal plus noise
Figure 2—Binary signal with additive noise at-6 db SNR
(a)
(b)
(c)
(d)
(e)
Figure 3—Adaption of the memory at-6 db SNR: (a) Blank initial memory; (b) Memory after first dump; (c) Memory after 12 dumps; (d) Memory after 40 dumps; (e) Perfect “checkerboard” memory for comparison
As the machine memory adapts to this noisy input signal, it progresses as shown in [Figure 3]. The sign of 103 memory components are displayed in a raster pattern in this figure. [Figure 3a] shows the memory in its blank initial state at the start of the adaption process. [Figure 3b] shows the memory after the first adaption of the memory. This first “dump” occurred after the threshold had decayed to the point where an energy measurement produced an acceptance decision. [Figure 3c] and 3d show the memory after 12 and 40 adaptions, respectively. These dumps, of course, are based on both energy and cross-correlation measurements. As can be seen, the adapted memory after 40 dumps is already quite close to the perfect memory shown by the “checkerboard” pattern of [Figure 3c].
The detailed analysis of the performance of this type of machine vs. signal-to-noise ratio, average signal repetition rate, signal duration, and machine parameters is extremely complex. Therefore, it is not appropriate here to detail the results of the analytical and experimental work on the performance of this machine. However, several conclusions of a general nature can be stated.
(a) Because the machine memory is always adapting, there is a relatively high penalty for “false alarms.” False alarms can destroy a perfect memory. Hence, the threshold level needs to be set appropriately high for the memory adaption. If one wishes to detect signal occurrences with more tolerance to false alarms, a separate comparator and threshold level should be used.
(b) The present machine structure, which allows for slowly varying changes in the signal waveshape, exhibits a marked threshold effect in steady-state performance at an input signal-to-noise ratio (peak signal power-to-average noise power ratio) of about -12 db. Below this signal level, the time required for convergence increases very rapidly with decreasing signal level. At higher SNR, convergence to noise-like signals, having good auto-correlation properties, occurs at a satisfactory rate.
A more detailed discussion of performance has been published in the report cited in footnote reference 1.
Conceptual Design of Self-Organizing Machines
P. A. Kleyn
Northrop Nortronics
Systems Support Department
Anaheim, California
Self-organization is defined and several examples which motivate this definition are presented. The significance of this definition is explored by comparison with the metrization problem discussed in the companion paper [(1)] and it is seen that self-organization requires decomposing the space representing the environment. In the absence of a priori knowledge of the environment, the self-organizing machine must resort to a sequence of projections on unit spheres to effect this decomposition. Such a sequence of projections can be provided by repeated use of a nilpotent projection operator (NPO). An analog computer mechanization of one such NPO is discussed and the signal processing behavior of the NPO is presented in detail using the Euclidean geometrical representation of the metrizable topology provided in the companion paper. Self-organizing systems using multiple NPO’s are discussed and current areas of research are identified.
INTRODUCTION
Unlike the companion paper which considers certain questions in depth, this paper presents a survey of the scope of our work in self-organizing systems and is not intended to be profound.
The approach we have followed may be called phenomenological ([Figure 1]). That is, the desired behavior (self-organization) was defined, represented mathematically, and a mechanism(s) required to yield the postulated behavior was synthesized using mathematical techniques. One advantage of this approach is that it avoids assumptions of uniqueness of the mechanism. Another advantage is that the desired behavior, which is after all the principal objective, is taken as invariant. An obvious disadvantage is the requirement for the aforementioned synthesis technique; fortunately in our case a sufficiently general technique had been developed by the author of the companion paper.
From the foregoing and from the definition of self-organization we employ ([see conceptual model]), it would appear that our research does not fit comfortably within any of the well publicized approaches to self-organization [(2)]. Philosophically, we lean toward viewpoints expressed by [Ashby (3)], [(4)], [Hawkins (5)], and [Mesarovic (6)] but with certain reservations. We have avoided the neural net approach partly because it is receiving considerable attention and also because the brain mechanism need not be the unique way to produce the desired behavior.
Figure 1—Approach used in Nortronics research on self-organizing systems
Nor have we followed the probability computer or statistical decision theory approach exemplified by [Braverman (7)] because these usually require some sort of preassigned coordinate system [(8)]. Neither will the reader find much indication of formal logic [(9)] or heuristic [(10)] programming. Instead, we view a self-organizing system more as a mirror whose appearance reflects the environment rather than its own intrinsic nature. With this viewpoint, a self-organizing system appears very flexible because it possesses few internal constraints which would tend to distort the reflection of the environment and hinder its ability to adapt.
CONCEPTUAL MODEL
Definition
A system is said to be self-organizing if, after observing the input and output of an unknown phenomenon (transfer relation), the system organizes itself into a simulation of the unknown phenomenon.
Implicit in this definition is the requirement that the self-organizing machine (SOM) not possess a preassigned coordinate system. In fact it is just this ability to acquire that coordinate system implicit in the input-output spaces which define the phenomenon that we designate as self-organization. Thus any a priori information programmed into the SOM by means of, for example, stored or wired programs, constrains the SOM and limits its ability to adapt. We do not mean to suggest that such preprogramming is not useful or desirable; merely that it is inconsistent with the requirement for self-organization. As shown in [Figure 2], it is the given portion of the environment which the SOM is to simulate, which via the defining end spaces, furnishes the SOM with all the data it needs to construct the coordinate system intrinsic to those spaces.
The motivation for requiring the ability to simulate as a feature of self-organization stems from the following examples.
Consider the operation of driving an automobile. [Figure 3] depicts the relation characterized by a set of inputs; steering, throttle, brakes, transmission, and a set of outputs; the trajectory. Operation of the automobile requires a device (SOM) which for a desired trajectory can furnish those inputs which realize the desired trajectory. In order to provide the proper inputs to the automobile, the SOM must contain a simulation of ⨍⁻¹(x).
Figure 2—Simulation of (a portion of) the environment
Figure 3—Simulation of a relation
Since ⨍(x) is completely defined in terms of the inputs and the resulting trajectories, exposure to them provide the SOM with all the information necessary to simulate ⨍⁻¹(x). And if the SOM possesses internal processes which cause rearrangement of the input-output relation of the SOM to correspond to ⨍⁻¹(x) in accordance with the observed data, the SOM can operate an automobile. It is this internal change which is implied by the term “self-organizing,” but note that the instructions which specify the desired organization have their source in the environment.
As a second example consider adaptation to the environment. Adapt (from Webster) means: “to change (oneself) so that one’s behavior, attitudes, etc., will conform to new or changed circumstances. Adaptation in biology means a change in structure, function or form that produces better adjustment to the environment.” These statements suggest a simulation because adjustment to the environment implies survival by exposing the organism to the beneficial rather than the inimical effects of the environment. If we represent the environment (or portion thereof) as a relation as shown in [Figure 2], we note that the ability to predict what effect a given disturbance will have is due to a simulation of the cause-effect relation which characterizes the environment.
It would be a mistake to infer from these examples that simulation preserves the appearance of the causes and effects which characterize a relation. We clarify this situation by examining a relation and its simulation.
Consider the relation between two mothers and their sons as pictured in [Figure 4]. Observe that if symbols (points) are substituted for the actual physical objects (mothers and sons), the relation is not altered in any way. This is what we mean by simulation and this is how a SOM simulates. It is not even necessary that the objects, used to display the relation, be defined; i.e., these objects may be primitive. (If this were not so, no mathematical or physical theory could model the environment.) The main prerequisite is sufficient resolution to distinguish the objects from each other.
Figure 4—A relation of objects—displayed and simulated
MATHEMATICAL MODEL
The mathematical model must represent both the environment and the SOM and for reasons given in the companion paper each is represented as a metrizable topology. For uniqueness we factor each space into equal parts and represent the environment as the channel
W ⟶ X. (Ref. 10a)
Consider now the SOM to be represented by the cascaded channels
X ⟶ Y ⟶ Z
where X ⟶ Y is a variable which represents the reorganization of the SOM existing input-output relation represented by Y ⟶ Z.
The solution of the three channels-in-cascade problem
