![]() |
Leon Cooper's somewhat peripatetic career has resulted in work in quantum field theory, superconductivity, the quantum theory of measurement as well as the mechanisms that underly learning and memory. He has written numerous essays on a variety of subjects as well as a highly regarded introduction to the ideas and methods of physics for non-physicists. Among the many accolades, he has received (some deserved) one he likes specially is the comment of an anonymous reviewer who characterized him as “a nonsense physicist”.
This compilation of papers presents the evolution of his thinking on mechanisms of learning, memory storage and higher brain function. The first half proceeds from early models of memory and synaptic plasticity to a concrete theory that has been put into detailed correspondence with experiment and leads to the very current exploration of the molecular basis for learning and memory storage. The second half outlines his efforts to investigate the properties of neural network systems and to explore to what extent they can be applied to real world problems.
In all this collection, hopefully, provides a coherent, no-nonsense, account of a line of research that leads to present investigations into the biological basis for learning and memory storage and the information processing and classification properties of neural systems.
Sample Chapter(s)
General Introduction (52 KB)
Part I. Physiological Basis of Learning and Memory Storage (533 KB)
https://doi.org/10.1142/9789812795885_fmatter
The following sections are included:
https://doi.org/10.1142/9789812795885_others01
Please refer to full text.
https://doi.org/10.1142/9789812795885_0001
A model of long-term memory, motivated by the anatomy and physiology of the mamalian central nervous system is proposed. We suggest that what is of importance to the nervous system is the collective individual activities of large numbers of individual neurons and that just such a collection of activities constitutes the memory trace. We assume that the long-term memory is the sum of such individual traces. With a minimum of some of the functions of a biological memory and that many of its properties are reminiscent of the brain. Throughout the emphasis is on consistency with the known physiology.
https://doi.org/10.1142/9789812795885_others02
Please refer to full text.
https://doi.org/10.1142/9789812795885_0002
A brief account is given of some of the properties of a neural model which displays on a primitive level features which suggest some of the mental behavior associated with animal memory and learning. The model as well as the basic passive procedure by which the neural network modifies itself with experience is consistent with known neurophysiology as well as with what information might be available in the neuron network. One must, however, assume that there exists a means of communication (electrical or chemical) between the cell body and dendrite ends–communication in a direction opposite to the flow of electrical signals. This same modification procedure could also lead to the formation of cells of the type observed by Hubel & Wiesel in the cat's visual cortex. It is suggestive that a network modification procedure that could produce such early processing cells might also be responsible, in cortex, for ‘higher’ mental processes. The explicit mathematics employed here is that of linear transformations on a vector space. However, as only certain topological properties are used, it is possible to construct a more general non-linear theory.
https://doi.org/10.1142/9789812795885_others03
Please refer to full text.
https://doi.org/10.1142/9789812795885_0003
Passive modification of the strength of synaptic junctions that results in the construction of internal mappings with some of the properties of memory is shown to lead to the development of Hubel-Wiesel type feature detectors in visual cortex. With such synaptic modification a cortical cell can become committed to an arbitrary but repeated external pattern, and thus fire every time the pattern is presented even if that cell has no genetic pre-disposition to respond to the particular pattern. The additional assumption of lateral inhibition between cortical cells severely limits the number of cells which respond to one pattern as well as the number of patterns that are picked up by a cell. The introduction of a simple neural mapping from the visual field to the lateral geniculate leads to an interaction between patterns which, combined with our assumptions above, seems to lead to a progression of patterns from column to column of the type observed by Hubel and Wiesel in monkey.
https://doi.org/10.1142/9789812795885_others04
Please refer to full text.
https://doi.org/10.1142/9789812795885_0004
We assume that between lateral geniculate and visual cortical cells there exist labile synapses that modify themselves in a new fashion called threshold passive modification and in addition, non-labile synapses that contain permanent information. In the theory which results there is an increase in the specificity of response of a cortical cell when it is exposed to stimuli due to normal patterned visual experience. Non-patterned input, such as might be expected when an animal is dark-reared or raised with eyelids sutured, results in a loss of specificity, with details depending on whether noise to labile and non-labile junctions is correlated. Specificity can sometimes be regained, however, with a return of input due to patterned vision. We propose that this provides a possible explanation of experimental results obtained by Imbert and Buisseret (1975); Blakemore and Van Sluyters (1975); Buisseret and Imbert (1976); and Frégnac and Imbert (1977, 1978).
https://doi.org/10.1142/9789812795885_others05
Please refer to full text.
https://doi.org/10.1142/9789812795885_0005
Please refer to full text.
https://doi.org/10.1142/9789812795885_others06
Please refer to full text.
https://doi.org/10.1142/9789812795885_0006
The development of stimulus selectivity in the primary sensory cortex of higher vertebrates is considered in a general mathematical framework. A synaptic evolution scheme of a new kind is proposed in which incoming patterns rather than converging afferents compete. The change in the efficacy of a given synapse depends not only on instantaneous pre- and postsynaptic activities but also on a slowly varying time-averaged value of the postsynaptic activity. Assuming an appropriate nonlinear form for this dependence, development of selectivity is obtained under quite general conditions on the sensory environment. One does not require nonlinearity of the neuron's integrative power nor does one need to assume any particular form for intracortical circuitry. This is first illustrated in simple cases, e.g., when the environment consists of only two different stimuli presented alternately in a random manner. The following formal statement then holds: the state of the system converges with probability 1 to points of maximum selectivity in the state space. We next consider the problem of early development of orientation selectivity and binocular interaction in primary visual cortex. Giving the environment an appropriate form, we obtain orientation tuning curves and ocular dominance comparable to what is observed in normally reared adult cats or monkeys. Simulations with binocular input and various types of normal or altered environments show good agreement with the relevant experimental data. Experiments are suggested that could test our theory further.
https://doi.org/10.1142/9789812795885_others07
Please refer to full text.
https://doi.org/10.1142/9789812795885_0007
A single-cell theory for the development of selectivity and ocular dominance in visual cortex has been generalized to incorporate more realistic neural networks that approximate the actual anatomy of small regions of cortex. In particular, we have analyzed a network consisting of excitatory and inhibitory cells, both of which may receive information from the lateral geniculate nucleus (LGN) and then interact through cortico–cortical synapses in a mean-field approximation. Our investigation of the evolution of a cell in this mean-field network indicates that many of the results on existence and stability of fixed points that have been obtained previously in the single-cell theory can be successfully generalized here. We can, in addition, make explicit further statements concerning the independent effects of excitatory and inhibitory neurons on selectivity and ocular dominance. For example, shutting off inhibitory cells lessens selectivity and alters ocular dominance (masked synapses). These inhibitory cells may be selective, but there is no theoretical necessity that they be so. Further, the intercortical inhibitory synapses do not have to be very responsive to visual experience. Most of the learning process can occur among the excitatory LGN–cortical synapses.
https://doi.org/10.1142/9789812795885_others08
Please refer to full text.
https://doi.org/10.1142/9789812795885_0008
Please refer to full text.
https://doi.org/10.1142/9789812795885_others09
Please refer to full text.
https://doi.org/10.1142/9789812795885_0009
The functional organization of the cerebral cortex is modified dramatically by sensory experience during early postnatal life. The basis for these modifications is a type of synaptic plasticity that may also contribute to some forms of adult learning. The question of how synapses modify according to experience has been approached by determining theoretically what is required of a modification mechanism to account for the available experimental data in the developing visual cortex. The resulting theory states precisely how certain variables might influence synaptic modifications. This insight has led to the development of a biologically plausible molecular model for synapse modification in the cerebral cortex.
https://doi.org/10.1142/9789812795885_others10
Please refer to full text.
https://doi.org/10.1142/9789812795885_0010
Please refer to full text.
https://doi.org/10.1142/9789812795885_others11
Please refer to full text.
https://doi.org/10.1142/9789812795885_0011
1. The aim of this work was to assess whether a form of synaptic modification based on the theory of Bienenstock, Cooper, and Munro (BCM) can, with a fixed set of parameters, reproduce both the kinetics and equilibrium states of experience-dependent modifications that have been observed experimentally in kitten striate cortex.
2. According to the BCM theory, the connection strength of excitatory geniculocortical synapses varies as the product of a measure of input activity (d) and a function (φ) of the summed postsynaptic response. For all postsynaptic responses greater than spontaneous but less than a critical value called the “modification threshold” (θ), φ has a negative value. For all postsynaptic responses greater than θ, φ has a positive value. A novel feature of the BCM theory is that the value of θ is not fixed, but rather “slides” as a nonlinear function of the average postsynaptic response.
3. This theory permits precise specification of theoretical equivalents of experimental situations, allowing detailed, quantitative comparisons of theory with experiment. Such comparisons were carried out here in a series of computer simulations.
4. Simulations are performed by presenting input to a model cortical neuron, calculating the summed postsynaptic response, and then changing the synaptic weights according to the BCM theory. This process is repeated until the synaptic weights reach an equilibrium state.
5. Two types of geniculocortical input are simulated: “pattern” and “noise.” Pattern input is assumed to correspond to the type of input that arises when a visual contour of a particular orientation is presented to the retina. This type of input is said to be “correlated” when the two sets of geniculocortical fibers relaying information from the two eyes convey the same patterns at the same time. Noise input is assumed to correspond to the type of input that arises in the absence of visual contours and, by definition, is uncorrelated.
6. By varying the types of input available to the two sets of geniculocortical synapses, we simulate the following types of visual experience: 1) normal binocular contour vision, 2) monocular deprivation, 3) reverse suture, 4) strabismus, 5) binocular deprivation, and 6) normal contour vision after a period of monocular deprivation.
7. The constraints placed on the set of parameters by each type of simulated visual environment, and the effects that such constraints have on the evolution of the synaptic weights, are investigated in detail.
8. It was discovered that the exact form and dependencies of the sliding modification threshold are critical in obtaining a set of simulations that are consistent with the experimentally observed kinetics of synaptic modification in visual cortex. In particular, to account for observed changes during reverse suture and binocular deprivation, the value of θ could approach zero only when the synaptic strengths were driven to very low values. In the present model, this was achieved by including in the calculation of θ the postsynaptic responses generated by spontaneous input activity.
9. It is concluded that the modification of excitatory geniculocortical synapses according to rules derived from the BCM theory can account for both the outcome and kinetics of experience-dependent synaptic plasticity in kitten striate cortex. The understanding that this theory provides should be useful for the design of neurophysiological experiments aimed at elucidating the molecular mechanisms in play during the modification of visual cortex by experience.
https://doi.org/10.1142/9789812795885_others12
Please refer to full text.
https://doi.org/10.1142/9789812795885_0012
In this paper, we present an objective function formulation of the Bienenstock, Cooper, and Munro (BCM) theory of visual cortical plasticity that permits us to demonstrate the connection between the unsupervised BCM learning procedure and various statistical methods, in particular, that of Projection Pursuit. This formulation provides a general method for stability analysis of the fixed points of the theory and enables us to analyze the behavior and the evolution of the network under various visual rearing conditions. It also allows comparison with many existing unsupervised methods. This model has been shown successful in various applications such as phoneme and 3D object recognition. We thus have the striking and possibly highly significant result that a biological neuron is performing a sophisticated statistical procedure.
https://doi.org/10.1142/9789812795885_others13
Please refer to full text.
https://doi.org/10.1142/9789812795885_0013
We tested a theoretical prediction that patterns of excitatory input activity that consistently fail to activate target neurons sufficiently to induce synaptic potentiation will instead cause a specific synaptic depression. To realize this situation experimentally, the Schaffer collateral projection to area CA1 in rat hippocampal slices was stimulated electrically at frequencies ranging from 0.5 to 50 Hz. Nine hundred pulses at 1–3 Hz consistently yielded a depression of the CA1 population excitatory postsynaptic potential that persisted without signs of recovery for >1 hr after cessation of the conditioning stimulation. This long-term depression was specific to the conditioned input, ruling out generalized changes in postsynaptic responsiveness or excitability. Three lines of evidence suggest that this effect is accounted for by a modification of synaptic effectiveness rather than damage to or fatigue of the stimulated inputs. First, the effect was dependent on the stimulation frequency; 900 pulses at 10 Hz caused no lasting change, and at 50 Hz a synaptic potentiation was usually observed. Second, the depressed synapses continued to support long-term potentiation in response to a high-frequency tetanus. Third, the effects of conditioning stimulation could be prevented by application of NMDA receptor antagonists. Thus, our data suggest that synaptic depression can be triggered by prolonged NMDA receptor activation that is below the threshold for inducing synaptic potentiation. We propose that this mechanism is important for the modifications of hippocampal response properties that underlie some forms of learning and memory.
https://doi.org/10.1142/9789812795885_others14
Please refer to full text.
https://doi.org/10.1142/9789812795885_0014
Activity-dependent synaptic plasticity in the superficial layers of juvenile cat and adult rat visual neocortex was compared with that in adult rat hippocampal field CA1. Stimulation of neocortical layer IV reliably induced synaptic long-term potentiation (LTP) and long-term depression (LTD) in layer III with precisely the same types of stimulation protocols that were effective in CA1. Neocortical LTP and LTD were specific to the conditioned pathway and, as in the hippocampus, were dependent on activation of N-methyl-D-aspartate receptors. These results provide strong support for the view that common principles may govern experience-dependent synaptic plasticity in CA1 and throughout the superficial layers of the mammalian neocortex.
https://doi.org/10.1142/9789812795885_others15
Please refer to full text.
https://doi.org/10.1142/9789812795885_0015
The Bienenstock, Cooper, and Munro (BCM) theory of synaptic plasticity has successfully reproduced the development of orientation selectivity and ocular dominance in kitten visual cortex in normal, as well as deprived, visual environments. To better compare the consequences of this theory with experiment, previous abstractions of the visual environment are replaced in this work by real visual images with retinal processing. The visual environment is represented by 24 gray-scale natural images that are shifted across retinal fields. In this environment, the BCM neuron develops receptive fields similar to the fields of simple cells found in kitten striate cortex. These fields display adjacent excitatory and inhibitory bands when tested with spot stimuli, orientation selectivity when tested with bar stimuli, and spatial-frequency selectivity when tested with sinusoidal gratings. In addition, their development in various deprived visual environments agrees with experimental results.
https://doi.org/10.1142/9789812795885_others16
Please refer to full text.
https://doi.org/10.1142/9789812795885_0016
In this paper we realistically model a two-eye visual environment and study its effect on single cell synaptic modification. In particular, we study the effect of image misalignment on receptive field formation after eye opening. We show that binocular misalignment effects PCA and BCM learning in different ways. For the BCM learning rule this misalignment is sufficient to produce varying degrees of ocular dominance, whereas for PCA learning binocular neurons emerge in every case. Such differences should help us distinguish between these learning rules.
https://doi.org/10.1142/9789812795885_others17
Please refer to full text.
https://doi.org/10.1142/9789812795885_0017
Please refer to full text.
https://doi.org/10.1142/9789812795885_others18
Please refer to full text.
https://doi.org/10.1142/9789812795885_0018
We present a general neural model for supervised learning of pattern categories which can resolve pattern classes separated by nonlinear, essentially arbitrary boundaries. The concept of a pattern class develops from storing in memory a limited number of class elements (prototypes). Associated with each prototype is a modifiable scalar weighting factor (λ) which effectively defines the threshold for categorization of an input with the class of the given prototype. Learning involves (1) commitment of prototypes to memory and (2) adjustment of the various λ factors to eliminate classification errors. In tests, the model ably defined classification boundaries that largely separated complicated pattern regions. We discuss the role which divisive inhibition might play in a possible implementation of the model by a network of neurons.
https://doi.org/10.1142/9789812795885_others19
Please refer to full text.
https://doi.org/10.1142/9789812795885_0019
The study of distributed memory systems has produced a number of models which work well in limited domains. However, until recently, the application of such systems to real-world problems has been difficult because of storage limitations, and their inherent architectural (and for serial simulation, computational) complexity. Recent development of memories with unrestricted storage capacity and economical feedforward architectures has opened the way to the application of such systems to complex pattern recognition problems. However, such problems are sometimes underspecified by the features which describe the environment is often non-specific. We will review current work on high density memory systems and their network implementations. We will discuss a general learning algorithm for such high density memories and review its application to separate point sets. Finally, we will introduce an distributions of non-separate point sets.
https://doi.org/10.1142/9789812795885_others20
Please refer to full text.
https://doi.org/10.1142/9789812795885_0020
Please refer to full text.
https://doi.org/10.1142/9789812795885_others21
Please refer to full text.
https://doi.org/10.1142/9789812795885_0021
We present a relaxation model for memory based on a generalized coulomb potential. The model has arbitrarily large storage capacity and, in addition, well-defined basins of attraction about stored memory states. The model is compared with the Hopfield relaxation model.
https://doi.org/10.1142/9789812795885_0022
The Coulomb Potential Learning (CPL) algorithm (Bachmann et al., 1987), which derives its name from its functional form's likeness to a coulomb charge potential, was originally motivated by the short-comings of the Perceptron (Rosenblatt, 1962) and the original Hopfield net (Hopfield, 1979). In the case of the Perceptron, it was clear almost from the outset that the linear-separability provided by the perceptron would not be sufficient to perform complex tasks. In the case of the original Hopfield model, the recall capacity of the network is low due. to non-orthogonal memories and the existence of spurious memories. The CPL algorithm addresses both of these problems by providing a simple network which is capable of both storing an arbitrarily large number of memories with perfect recall and no spurious memories; and also constructing arbitrary nonlinear boundaries for classification tasks. In addition the CPL algorithm is easy to implement in hardware and is readily adaptable to parallel computation.
https://doi.org/10.1142/9789812795885_others22
Please refer to full text.
https://doi.org/10.1142/9789812795885_0023
Please refer to full text.
https://doi.org/10.1142/9789812795885_others23
Please refer to full text.
https://doi.org/10.1142/9789812795885_0024
Please refer to full text.
https://doi.org/10.1142/9789812795885_others24
Please refer to full text.
https://doi.org/10.1142/9789812795885_0025
Please refer to full text.
https://doi.org/10.1142/9789812795885_0026
This paper extends our previous results for averaged regression estimators (Perrone and Cooper, 1993b). We prove that averaged regression estimates always perform as good or better that their unaveraged counterparts. We show that this result derives from the notion of convexity and in this way demonstrate that a wide variety of optimization algorithms can benefit from averaging including: Mean Square Error, a general class of LP-norm cost functions, Maximum Likelihood Estimation, Maximum Entropy, Maximum Mutual Information, the Kullback-Leibler Information, Penalized Maximum Likelihood Estimation and Smoothing Splines.
https://doi.org/10.1142/9789812795885_others25
Please refer to full text.
https://doi.org/10.1142/9789812795885_0027
An unsupervised neural network model inductively acquires the ability to distinguish categorically the stop consonants of English, in a manner consistent with prenatal and early postnatal auditory experience, and without reference to any specialized knowledge of linguistic structure or the properties of speech. This argues against the common assumption that linguistic knowledge, and speech perception in particular, cannot be learned and must therefore be innately specified.
https://doi.org/10.1142/9789812795885_others26
Please refer to full text.
https://doi.org/10.1142/9789812795885_0028
In this paper we present a new version of the standard multilayer perceptron (MLP) algorithm for the state-of-the-art in neural network VLSI implementations: the Intel Ni1000. This approach enables the standard MLP to utilize the parallel architecture of the Ni1000 to achieve on the order of 40000, 256-dimensional classifications per second. Due to the compact size and affordable price of the Ni1000, this classification speed could be available for the average personal computer.
https://doi.org/10.1142/9789812795885_others27
Please refer to full text.
https://doi.org/10.1142/9789812795885_0029
Please refer to full text.
https://doi.org/10.1142/9789812795885_bmatter
The following sections are included:
Leon Cooper was born in 1930 in New York where he attended Columbia University (AB 1951: AM 1953: PhD 1954). He became a member of the Institute for Advanced Study (1954–55) after which he was a research associate at the University of Illinois (1955–57) and later an assistant professor at the Ohio State University (1957–58). He joined Brown University in 1958 where he became Henry Ledyard Goddard University Professor (1966–74) and where he is presently the Thomas J W Watson, Sr. Professor of Science (1974–) and Director of the Institute for Brain and Neural Systems.
In 1972 he received the Nobel Prize in Physics (with J Bardeen and J R Schrieffer) for his work on the theory of superconductivity which was completed while still in his 20s. In 1968, he was awarded the Comstock Prize (with J R Schrieffer) of the National Academy of Sciences. The Award of Excellence, Graduate Faculties Alumni of Columbia University and Descartes Medal, Academie de Paris, Universite Rene Descartes were conferred on Professor Cooper in the mid 1970s. In 1985 Professor Cooper received the John Jay Award of Columbia College and in 1990 the Columbia University award for Distinguished Achievement.