Professional Documents
Culture Documents
Patterns of activities of neurons serve as attractors, since they are those neuronal configurations which
correspond to minimal ’free energy’ of the whole system. Namely, they realize maximal possible agreement
among constitutive neurons and are most-strongly correlated with some environmental pattern. Neuronal
patterns-qua-attractors have both a material and a virtual aspect. As neuronal patterns, on the one hand,
patterns-qua-attractors are explicit carriers of informational contents. As attractors, on the other hand,
patterns-qua-attractors are implicit mental representations which acquire a meaning in contextual relations
to other possible patterns.
Recognition of an external pattern is explained as a (re)construction of the pattern which is the most
relevant and similar to a given environmental pattern. The identity of the processes of pattern construction,
re-construction and Hebbian short-term storage is realized in a net.
Perceptual processes are here modeled using Kohonen’s topology-preserving feature mapping onto cortex
where further associative processing is continued. To model stratification of associative processing because
of influence from higher brain areas, Haken’s multi-level synergetic network is found to be appropriate.
The hierarchy of brain processes is of ”software”-type, i.e. virtual, as well as it is of ”hardware”-type, i.e.
physiological. It is shown that synergetic and attractor dynamics can characterize not only neural networks,
but also underlying quantum networks. Neural nets are alone not sufficient for consciousness, but interaction
with the quantum level might provide effects necessary for consciousness, like, for instance, ultimate binding
of perceptual features into an unified experience.
It is mathematically demonstrated that associative neural networks realize information processing analogous
to the quantum dynamics. Parallels in the formalism of neural models and quantum theory are listed.
Basic elements of the quantum versus neural system (modeled by formal neurons and connections) are very
different, but their collective processes obey similar laws. Specifically, it is shown that neuron’s weighted
spatio-temporal integration of signals corresponds to the Feynman’s version of the Schrödinger equation.
In the first case weights are synaptic strengths determined by the Hebb or delta correlation rule; in the
second case weights are Green functions or density matrices. In both cases encodings of pattern-correlations
represent memory. (Re)construction of a neuronal pattern-qua-attractor is analogous to the ”wave-function
collapse”. Transformations of memory (or sub-conscious) representations to a conscious representation is
modeled in the same way.
Found mathematical analogies allow translation of the neural-net ”algorithm”, which in author’s simulations
works very well, into a quantum one. This indicates how such quantum networks, which might be exploited
by the sub-cellular levels of brain, could process information efficiently and also make it conscious.
Key words: neural net, quantum, brain, associative, synergetic, perception, consciousness
PACS numbers: 87.18.Sn, 87.19.La, 03.67.Lx, 89.70.+c, 87.18.Hf
Overall introduction and Two present the topic in a new style, but the
third part is a purely original contribution which
This paper is a review of models of multi-level brain indicates quantum neural-net-like information pro-
processing with connectionist approach. Parts One cessing and their potential relevance for conscious
perception. The aim is that all parts together model, order parameters, which can be embodied
should present an integrated overview of modeling by cardinal cells or cardinal domains, act as sym-
brain synergetics — at least from a particular (au- bolic codes or representatives of patterns.
thor’s) viewpoint, although not exclusively. Quantum processing will be argued to be neces-
In the first part a comprehensive introduction to sary for consciousness which remained absent from
models of associative neural networks is presented ordinary neural-net processing. Reasons for this
as well as their role in large-scale modeling of micro- will be listed. Then, neural-net mathematics and
cognition. Patterns of neuronal activity are inter- quantum mathematics will be systematically com-
preted as attractors of the parallel-distributed dy- pared. The resulting neuro-quantum mathematical
namics of an underlying complex dynamical system. parallelism, which shows an analogy or resemblance,
It will be shown how the system of synaptic not any rigorous and isomorphic equivalence, in-
connections represents memory. The transmission dicates that the network metaphor can be used for
rates of individual synapses represent correlations quantum as well as neural parallel-distributed pro-
between neurons, whereby each neuron may coop- cessing, in spite of differences of the nature of neural
erate in global encoding of many patterns as pre- and quantum ”units”.
scribed by the Hebb learning rule. Learning is a If we use these neuro-quantum analogies, effec-
process of adjusting synaptic strengths in order to tive information processing, similar to neural-net
store essential information. processing, can be searched for in quantum, sub-
In the second part processes of categorization, cellular and other synergetic systems, because the
adaptation, data compression and abstraction, fil- principles of collective information processing are
tering of relevant information and extraction of already known from our (including author’s) com-
novel information from stimuli, and perceptual pre- puter simulations of neural nets.
processing of patterns will be presented (now more
in detail) using the Kohonen model of topologically-
correct feature mapping and its extensions. This 1 Introduction to global brain
model is important for modeling perception and mo- processing: Associative or at-
tor action. Instead of free-energy minimization of
the previously-presented Hopfield model, in Koho- tractor neural networks
nen’s model these tasks are resolved by error mini-
mization, lateral inhibition, and winning of one neu- 1.1 Introduction to neural networks
ron. This is the cardinal neuron which individu- The term neural networks has several meanings. For
ally encodes a whole stimulus-pattern in the recep- neurobiologists it denotes webs of neural cells (neu-
tive field. Topology-preserving mapping means that rons) in the brain [20, 83]. Neurons are connected
similar input data are projected into adjacent car- through synapses which are essential for encoding
dinal neurons. This self-organizing processes is re- of the memory-traces. Information is processed by
sponsible for the formation of brain maps with lo- collective dynamics of neurons and synapses.
calized encoding of perceptual data. There, even Interdisciplinary oriented scientists started to
more, correlated patterns are stored next to each model nets of neurons and synapses mathematically
other. Methods of vector quantization and princi- in the fourties [6]. Since then, the term neural net-
pal component analysis will also be presented. works is usually a short name for neural network
In the third part various levels of synergetic com- models. Another term for neural network mod-
putation in the cortex and their relations will be els, which were realized by computer simulations or
presented. Emergence of a hierarchy of neuronal were implemented in hardware, is artificial neural
patterns acting as attractors, and occurrence of networks. The spectrum of neural network models
thought processes as dynamic pattern sequences or spreads from biologically-plausible models of real
episodes will be discussed. In Haken’s synergetic neural processing in living systems on one side to
technical imitations applied for information process- 1. Fully and bidirectionally connected associative
ing on the other side. The second branch includes or attractor neural networks recognize and store pat-
all simple artificial brain-like mechanisms or algo- terns without reducing information content. Pat-
rithms which establish a sort of model of external terns are stored completely, e.g. as pixel images,
data-structures [12, 43, 99]. This is called empirical representing attractors of the system dynamics.
(external) modeling of physical, chemical, techno- Neurons are symmetrically or homogeneously dis-
logical, economical and other processes. tributed (figure 3 left).
Here we will be concerned with the first branch 2. Uni-directionally oriented networks with ”hidden
of neural network models — the theoretical (inter- units” (cardinal neurons or order-parameter-cells)
nal) models of neural network architectures [27, 89]. which are located in the middle layer. This is the
This kind of analysis has two sub-branches: model- second layer of hierarchically ordered population of
ing, as faithful as possible, of real neural webs, and competitive neurons (figure 6). Such networks are
mathematically-oriented research of neural network usually feed-forward nets (multi-layer perceptrons)
models. Of course, it is often not possible to make a [18, 58] which sometimes have feed-back loops exe-
sharp distinction between these branches and sub- cuting the so-called back-propagation of errors.
branches of modeling. The mathematical analysis Neural networks are currently the most successful
is more or less developed from neurophysiological model in cognitive neuroscience [79, 67]. Associative
data. In some cases these data are insufficient, but neural network models [12, 54, 56, 40, 72, 27, 36, 89]
in other cases models are insufficient, because they are a suitable mathematical description and simu-
only very roughly obey biological data. lation of the processing in associative areas of the
Mathematical approaches to neural network cortex [20, 83].
models used for theoretical or internal modeling The large-scale dynamics of an associative neural
(”physics of neural networks”) can be classified as net can be represented in the configuration-energy
follows: space (figure 1). Each point on the horizontal axis
1. Haken’s synergetic computers [40]; represents a unique configuration, denoted by vector
2. statistical mechanics of neural networks which ~q. Points on the vertical axis represent the potential
was developed from the Hopfield model and spin or free-energy function E of each such configuration
glass theory [45, 102, 3]; ~q [45]. The vector ~q = (q1 , q2 , ..., qN ) specifies the
3. self-organizing feature maps by Kohonen and state of the entire neural net, i.e., the entire neu-
others [56, 54]. ronal configuration. Each component qi describes
exactly the state of each constituent neuron of this
In this introductory part we examine neural net- configuration.
work models, henceforth referred to simply as ’neu- As can be seen in figure 1, the current neuronal
ral networks’, which are the most promising and il- configurations, represented by a black ball, travels
lustrative ones for micro-cognitive modeling [67, 79]. in the configuration-energy space of possible config-
In modeling we usually neglect whole internal struc- urations in order to find a stable state. Where the
ture of an individual neuron and an individual ball is stabilized, configuration represents a pattern.
synapse. We assume that all information processing
is a result of collective dynamics of the whole system
of neurons and synapses. Each neuron is receiving
1.2 Basic neurophysiological facts and
signals from numerous other neurons. This is mod- their modeling
eled by a simple addition of such synapse-weighted The body of a neuron, the soma, acts as an ’adder’
contributions reflected in an increase of neuron’s ac- in that it receives signals from all other neurons and
tivity. adds them together [46]. Before being added, a sig-
Due to their architecture neural network models nal must be multiplied in the soma by the strength
could be divided into following groups: of the synaptic connection reflecting both its trans-
all for one’. In natural neural nets each neuron is not that bottom pattern configuration [3]. This implies
necessarily interconnected directly with all others. that neurons fire or are quiescent in such a way as
However, the network behaves as though all its neu- to form patterns. Once the system ’falls into’ such
rons were interconnected in this way (mathematico- a configuration, all further configuration-changing
physical analysis in: [70](pp.45-48) and [72]). A processes cease until the reception of a new stimu-
large number of neurons is densely interconnected lus.
with a large number of other neurons. In the ab- This process is depicted in figure 1. The state
sence of direct connections, connections are estab- of an individual neuron is represented as a pixel of
lished via a mediator. In either case, the essential the picture. The transformations from a configura-
property of such systems of neurons is that neu- tion which resembles nothing recognizable to that
rons being closely inter-related via a system of con- of a pattern (’spiral’) is readily apparent. Paral-
nections one to all others. Therefore various forms lel to configurations the corresponding changes to
of collective organization of neuronal states occur configuration-energy space are drawn. Retrieval of
which give rise in turn to the formation of ’at- another pattern (’E’) from memory corresponds to
tractors’ or ’potential wells’ in configuration-energy converging to another attractor. It is critical to
space. Configurations with potential wells or at- know that the energy surface changes in the pro-
tractors are called ’patterns’. It is exactly such cess! While recognizing the first pattern, the first
patterns-qua-attractors which are major factors in potential well is especially deepened; while recalling
determining the course of mental processes, not the the second pattern, the second attractor becomes
neurons and synapses alone. Therefore even large- dominant and larger. Thus, it is obvious that a
scale lesions don’t destroy functioning essentially if neural network is a very dynamic system in which
attractor structures are preserved. continual change of configuration-energy space, i.e.,
Let us consider the case of a network of receiving changing the configurations of the system, enables
neurons (sensors or receptors) in an eye. The net re- flexible formation and storage of patterns.
ceives a stimulus from a certain environment. Neu-
rons thus stimulated begin to organize themselves.
The self-organizing process is then transferred to
1.4 Recognition and memorizing of pat-
centers of higher brain function, firstly to the cen- terns
ters for visual perception in the cerebral cortex. Figure 2 illustrates once again the process of
Newly-formed patterns in the cerebral cortex inter- pattern-formation and/or pattern-reconstruction
act associatively with other patterns... The process (recognition). We may conceive of configuration-
of neuronal organization consists, in part, in switch energy space (figure 1) as a ’soft’, deformable sur-
of neuronal states on a massive scale. Synaptic con- face of a mattress with shallow pits and the momen-
nections are being strengthened simultaneously. tary configuration of the network as a heavy ball
The process of changing of synaptic strengths resting conformally on this surface. The ball deep-
in response to various stimuli is called learn- ens the potential well centered around that portion
ing. During such processes neuronal configurations of configuration-energy space where the ball is lo-
and synaptic connections, which represent memory- cated at the given moment. External stimuli are
components, undergo change. Such processes occur “putting” the ball, which represents the system, in
always in such a way that the energy of the entire a certain configuration, i.e., they locate the initial
system decreases. The bottom of a potential well point in configuration-energy space. These stimuli
in such a configuration represents a pattern and at- force the receptor neurons into a configuration of
tracts, in turn, all other configurations. Therefore, stimuli (which is none other than the distributed
a point at the bottom of a potential well is called an representation of the external stimuli themselves)
attractor: all configurations within the boundaries which propagate, in turn, to regions of higher men-
of a potential well in configuration space converge to tal function in the cerebral cortex.
mostly on bio-electrical processes involving neurons, lations between neuronal activities of a single pat-
whereas long-term storage depends on bio-chemical tern.
changes in synapses. Each individual product in equation (2) repre-
sents coupling. If an individual product is negative
1.5 Hebbian learning based on search of (positive), then connection is weakened (strength-
correlations ened). The connection strength Jij is determined by
the sum of such coupling of all patterns which are
A neural network can store many patterns simulta- stored simultaneously in the same system. There
neously. Accordingly, each neuron and each synapse are P of stored patterns. Furthermore, if the states
can play multiple roles, since each neuron (or each of two connected neurons vik and vjk collaborating
synapse) participates in several patterns. There can
in the formation of the k th neuron are of the same
be several potential wells in configuration-potential
sign, then they will contribute to the stability of the
space, to which the dynamics of the system con-
k th pattern in the neural network. If their signs are
verges. However, only one pattern ~v is realized in
different, then they will weaken the pattern.
the system of neurons ~q at a moment. During the
Patterns stored in synaptic connections are
pattern-recall the following equality holds: vik = qi .
loaded one atop another. Connections are deter-
The strength of synaptic connections are deter-
mined locally according to mutual (dis)agreement of
mined by Hebb’s ’learning rule’ which is a fact of
pairs of neurons. Globally, in turn, connections de-
neurophysiology [6]: If two neurons are both active
termine the (in)stability of all configurations. Thus,
or both inactive, then the synaptic connection be-
neurons locally compete amongst themselves, each
tween them is strengthened. Otherwise, if one is
seeking to prevail and convert the rest to its state.
active and the other is inactive, then their mutual
A neuron succeeds in this if it is in best agreement
synaptic connection is weakened. Therefore, if two
with the other neurons and if it has the most sup-
neurons corroborate each other, their connection is
port from the rest of the system. Then the neuron
strengthened. If, however, they contradict or coun-
has excitatory connections.
teract each other, then their connection is weak-
Globally, neurons mediate competition and cou-
ened.
pling among entire patterns. That’s similar to the
Let vik denote a variable ranging over patterns,
case of two individual sportsmen (“neurons”) which
where k refers to the k th pattern and i to the ith
mediate in an international sporting event compe-
neuron. So, vik represents the role which the ith
tition of their respective states (“patterns of neu-
neuron plays in forming the k th pattern. Now, let
rons”) which they are representing at the event,
us formulate Hebb’s rule mathematically:
thus bringing about a ’meeting of two states’.
If both the ith and j th neurons are either both
In figure 3, on the left, a ’natural’ representation
active in forming the k th pattern (vik = vjk = 1)
of Hopfield’s network is depicted and, on the right,
or both are inactive in forming the k th pattern a diagram of the functional scheme of the structure
(vik = vjk = −1), then their mutual connection i − j of the memory or correlation matrix J according to
becomes stronger: Jij is increased. Otherwise, if Hebb’s equation (2). Individual elements Jij repre-
their activity-rate differs (vik = −1 and vjk = 1) sent the coupling of the ith neuron vik and the j th
or (vik = 1 and vjk = −1) then their mutual con- neuron vjk , forming parts of a given k th pattern.
nection becomes weaker. As can be shown, Hebb’s
auto-correlation equation, or Hebb’s learning rule,
consolidates all these facts: 1.6 Content-addressable memory in net-
P
works
X
Jij = vik vjk . (2) In a neural network information is stored in a
k=1
parallel-distributed (non-local) manner in the val-
It describes how synaptic connections encode corre- ues of connections: patterns are distributed across
patterns will not be stable and may move around in old, specifically, its own threshold or its threshold
configuration space. This means that the potential as determined by others.
well moves or (what is essentially the same thing) The Hebb learning rule gives such a value to J
it rolls along configuration space from one pattern that E is automatically minimized if stored patterns
into another. One pattern disappears and another ~v enter into the system of neurons ~q, i.e., ~q = ~v .
appears. The system may describe periodic (after a Only in that case the whole network is, with respect
certain time period the system returns to its orig- to the system of neurons as well as the system of
inal configuration), quasi-periodic, or completely-
connections, in an optimal overall agreement, and
chaotic paths. Dynamical pattern sequences are the
its free energy E is minimal.
origin of the flow of thought.
Patterns could not appear on the basis of sim-
ple data, which (sensory) neurons receive repeatedly
1.7 Explicit/implicit nature of patterns/
and in fragments, if these patterns would not collab-
attractors; cardinal cells; minimizing
orate and adapt to one-another. On a higher level,
free energy
a pattern could not collaborate in the formation of
Let us examine how neurons attain the complex patterns, if it weren’t in a comparative re-
energetically-optimal state. This is the config- lation with other patterns. The pattern would be
uration in which they are in optimal mutual meaningless by itself.
agreement. Free energy is proportional to the So, a pattern has a role only in relation to other
degree of disequilibrium of or disagreement among patterns and usual configurations. Its place in a set
a population of neurons. The optimal state can of relations is evident in the energy-configuration
only be attained through continual transactions space. A pattern has more meaning than an or-
through connections. dinary configuration, since it is more stable. It is
Energy ratios of configurations are carriers of more stable, because it is a more accurate represen-
implicit structures – attractors. The free-energy- tation of an actual situation - in the outside world
function E of the system is lower if neurons agree or on a more abstract internal level. Stated other-
(their activities qi and qj have the same sign in that wise, a pattern is more stable, since it agrees with
particular moment) while having positive connec- outside patterns and with other internal patterns.
tion Jij (good history of agreement) or disagree This is expressed in the fact that the free energy
while having negative connection (more disagree- of a system of neurons is lower when that pattern
ments in previous times). Furthermore, E increases is reconstructed in the system’s configuration. It is
if neurons agree while having negative synapse or because of this that a network tends to this pattern.
disagree while having positive synapse. In the same Let us emphasize again that a map from the envi-
time the energy is inversely-proportional to the ronment ’impresses’ contents into patterns, whereas
agreement of neurons with the neuronal field (“pub- their meaning emerges from their relations to one
lic opinion”). These conditions are encoded in the another and to the environment. External patterns
following equation [36]: from the environment have the same role as do inter-
nal patterns except that they have greater weight.
N X N N
1X 1X In figure 4 two configurations (denoted by a ball)
E=− Jij qi qj − Ti qi . (3)
2 i=1 j=1 2 i=1 are shown which are in different ratios to the other
configurations (other points on the horizontal axis).
Ti represents the influence of the neuronal field on The left one is not stable and the system assumes it
the ith neuron (the influence of “society” on the only transitionally, since the relief just transformed
individual, or the other way around). In another itself and the system hasn’t yet adapted to the
interpretation, Ti may represent a neuron’s thresh-
FIG. 4. The same location in the configuration-energy FIG. 5. A pattern (e.g., ~v k=2 ) is represented implic-
space (the ball) has various meanings in various contexts. itly when the system is in its neighborhood of attrac-
On the left it represents a pattern, on the right an ordi- tion (left: unstable initial situation). When the system
nary configuration. reaches the bottom 2 (right: stable final situation), the
pattern (black - white - black - white - black - white) be-
Minima of configuration-potential space repre- comes explicitly manifested. This may be accompanied
by the firing of a cardinal neuron ck=2 . (Note labels:
sent implicit order in a neural network and this
1, 3 – local spurious attractor; 2 – global attractor; 4 –
order determines the formation of further virtual basin of attraction)
mental structures. In figure 5, on the left, the sys-
tem is shown in the neighborhood of an attractor
and is certainly going to move towards a minimum the pattern is represented implicitly in the system
in configuration space, except if the surface of the of neurons, not only in memory (in the system of
relief would suddenly change because of other influ- synaptic connections). When the system of neu-
ences. In this case, the pattern is already latently rons reaches its attractor and, thereby realizing this
present in the system of neurons. It is contained pattern, in the same time, the value of its cardinal
implicitly in the state vector ~q, although it is mixed neuron attains a maximum, if it can attain various
together with traces of other patterns. Then, when degrees of strength or intensity. If it can only be
the system goes over to the bottom of the poten- active or inactive, then it will be activated at that
tial well (figure 5, on the right), a certain pattern time. With this, the cardinal neuron signifies that
appears clearly and resolutely ’in consciousness’. there is a corresponding pattern ’in consciousness’.
Then the pattern is manifest explicitly. This ex- In sending signals throughout the brain, the cardi-
plicit manifestation is especially obvious if the cor- nal neuron informs of the presence of its pattern and
responding cardinal neuron is selected and starts to represents it. Therefore a pattern (which is parallel-
fire. Switching-on of some most dominant neuron, distributed throughout a network at a lower level of
which then represents the whole neuronal pattern, its hierarchical structure) is encoded implicitly in the
often accompanies the explication of a pattern from concrete material state of one cell.
background. Since the moment of his winning the The use of cardinal neurons is but a special way
competition, such a cardinal neuron encodes and of organizing the operation of a neural network.
represents its pattern at the next hierachical level This principle obtains on the virtual level as well,
[40, 73, 75]. on which the cardinal neurons are replaced by syn-
When the system is in the basin of attraction ergetic order parameters which “act” implicitly.
of an attractor corresponding to a pattern, then There is a continual transition of configurations
The two-dimensional maps are mostly located in ally represent synaptic connections between the
the primary layer of the neocortex. Examples are potentially-cardinal neuron at location ~r and the
somatotopic map (in somatosensory cortex) of the sensory cells with index l.
skin surface, retinotopic map (in visual cortex) of
the eye’s retina, tonotopic map (in auditory cor-
tex) of the ear’s cochlea-spiral, etc. In the auditory-
cortex-map pitch (frequency) is encoded in one di-
rection, amplitude in the orthogonal direction. In
superior colliculus in midbrain direction and ampli-
tude of rapid eye movements (saccades) are spatially
encoded [82]. Triggering mechanism for flexion and
extension movements of muscles are encoded in the
motor cortex (ring map).
In the neocortex there are, vertically cutting
its layers, the cortical columns and micro-columns.
They can also be understood as a product of self-
organized topology-conserving mapping. They are FIG. 9. Kohonen network presented by a block of synap-
thus specialized to analyze some particular features tic strengths or synaptic weights. Each sensory cell xl is
of the input-patterns (orientations, velocities and connected to each neuron in the 7 x 7 output-map. 6-
dimensional input ~x (here presented one only: p = 1) is
directions of movements, edges, periodicity, color
projected into 2-dimensional map of 7 x 7 neurons (as
shades, etc.), or to regulate motor actions (e.g., many as there are squares on the top). Each neuron is
innervation of a common muscle). In association presented by a vertical column which corresponds to the
areas, columns are densely connected to execute vector of synapses between this neuron and all sensory
collectively complex tasks like recognition of faces, cells xl (l = 1, ..., 6). A “level” of synaptic-weight-values
speech comprehension and planning of arm trajecto- which are given by the 3rd input variable x3 is shaded.
Dark-shaded circles correspond to the 3rd synapse on the
ries. Columns provide informational basis for higher
neuron (1,1), and the 3rd synapse on the neuron (6,5). In
brain functions modeled by attractor neural net- this network-block 49 uncorrelated 6-dimensional input-
works. patterns could simultaneously be stored. If inputs are
correlated, memory compression starts... (From Zupan
[103], with permission.)
2.3 Learning process by error minimiza-
tion
If some prototype w ~ ~r matches the input-pattern
Using topology-preserving mapping a large set of p ~xk more than other prototypes do, then the cor-
input-patterns ~xk (k = 1, ..., p) gets encoded into a responding potentially-cardinal neuron at location
smaller set of p0 prototypes w ~ ~r . Here ~r is used as an ~r wins the competition against other potentially-
index. It refers to location of those neuron to which cardinal neurons ~r0 and becomes actually-cardinal
synapses are transmitting signals from sensory cells. neuron – the winner. In this case we say that the
The k th input-pattern is described by a vector ~xk = network has recognized the input pattern ~xk so that
(xk1 , xk2 , ..., xkn ). Its components represent individual it classified this input as an example of the pro-
pixels of the input-pattern which are projected into totype w ~ ~r . By meeting every new input-pattern
activities of individual sensory cells on the retina the prototype changes somewhat: the new updated
(vision), cochlea (audition), etc. n is the number of prototype is a result of a compromise between the
sensory cells. old prototype and new input. The prototype is en-
The prototype is represented by a vector w ~ ~r = coded in the synaptic connections w~rl and/or in the
(w~r1 , w~r2 , ..., w~rn ). Its components w~rl actu- cardinal neuron at location ~r. We cannot localize
the encoding since the synaptic strengths w~rl and for neighboring neurons, and J~r~r0 < 0 (inhibitory
the resulting activity of the cardinal neuron ~r both synapses) for neurons more far away. So, around
contribute to the mapping as a dynamic process. each neuron there is first a concentric circle of pos-
Statics does not mean anything to the network; itive synapses (support of similarly-oriented neu-
only dynamics has meaning in the sense of realiz- rons), and then a concentric circle of negative
ing specific input-output transformations. There- synapses (famous lateral inhibition, found in bio-
fore mathematical description which is static is not logical perceptual systems). Those neuron who has
enough for understanding neural network process- the most strong support of the nearest neurons (its
ing. One needs a lot of intuitive, flexible “mental inside domain), and is at the same time able to in-
modeling” also. hibit neurons in the neighboring domains (outside
All equations written from now on will be valid its domain), will become the winner. The concen-
for all sequentially occurring adaptation steps fol- tric order of synaptic strengths is illustrated by the
lowing each arrival of an input-pattern ~xk , therefore ”Mexican-hat” function (figure 10).
the index k will always be omitted. The activity of
some potentially-cardinal neuron in the second layer
(target-layer or map) is given by [82]
n
X X
c~r = f ( w~rl xl + J~r~r0 c~r0 − T~r ). (4)
l=1 r0
~
In the Kohonen model it is usual to represent the the difference between the individual input-stimuli
target-pattern by a vector of strengths of synap- xl and target-activities w~rl is bigger, or that the
tic connections (w ~ ~r ) from each sensory cell l (l = rates of synaptic transmission xl should be changed
1, ..., n) to a given neuron ~r. By this definition less if the difference is smaller. So, larger disagree-
input-pattern ~x and target-pattern w ~ ~r have the ment requires more adaptation. This is not all: the
same dimension (n) and can be subtracted. Now spatial distance between neurons is also important.
we see what is advantage of such kind of math- Its role is encoded in the shape of h~r~r0 which controls
ematical description. Activities of neurons c~r are the size of the neighborhood region. With that the
not important in this model. Only their relations to number of neurons affected by a single adaptation
configurations of sensory-cells are important. These step (7) is determined, as well as the rate of affec-
relations are encoded in configurations of synaptic tion. The target-patterns w ~ ~r associated with the
connections (w ~ ~r ) to a given neuron (~r). It is enough neighboring potentially-cardinal neurons ~r are cou-
if a neuron is denoted by its location ~r only. pled more strongly than target-patterns w ~ ~r corre-
The version (5) chooses the most excited neuron sponding to more distant potentially-cardinal neu-
to be the winner. The version (6) chooses those rons ~r .
cardinal neuron which represents the target-pattern Learning is the process where the adjustments
closest to the input-pattern ~x. Both ways bring the (5) or (6), respectively, and the adjustment (7)
same effect. are following each other in an iteration-procedure
The strengths of synaptic connections are [72, 82, 12]. Equations (5), (6) and (7) are equally
changed according to valid for all adaptation steps marked by k if we
keep in mind that ~xk is new in each step. In the
∆w~rl = ²h~r~r0 (xl − w~rl ) beginning of each (k th ) cycle of iteration the new
or ∆w
~ ~r = ²h~r~r0 (~x − w
~ ~r ) (7) input-pattern ~xk is presented to the network. Then
the winning neuron is chosen by (5) or (6), and
where synapses are changed according to the rate of sim-
à !
−(~r − ~r0 )2 ilarity between the input-pattern and the target-
h~r~r0 = exp . (8) pattern. In a model it can be prescribed whether
2σ 2
only one learning cycle or many cycles are follow-
σ is the radius of the Gaussian excitation function ing each input. Self-organizing adaptation causes
h~r~r0 . σ is fixed in the first approximation, but in a shift of the input-vector ~xk and the target-vector
the second approximation it can be slowly decreased w
~ ~r closer to one another. The adaptation process
during learning. In equation (8) ~r0 marks the ex- is finished when w ~ ~r = ~xk . We say that in this mo-
citation maximum gained by the actually-cardinal ment the input-pattern was recognized and simulta-
neuron. The “shape” of h~r~r0 is fixed with respect neously it gets stored in the new vector of synapses
to variation of ~r (locations of the neurons in sur- w
~ ~r . In such a way we get a new target-vector which
roundings). The position of Gaussian function h~r~r0 is a new memory representation. If the network is
is chosen by an incoming stimulus which enables already filled with old memories, we at least get a
some specific potentially-cardinal neuron to become better memory representation, because it considers
the actually-cardinal neuron ~r0 . The response of also the last input-data.
nearby neurons ~r is given by h~r~r0 . Introduction of The learning rate ² can also decrease gradually
this function is necessary in order to achieve local- with the number (k) of learning or adaptation steps.
ized neuronal excitation as in figure 8. Without proof we shall mention that two important
The prescription (7) originates from neurophysi- conditions are necessary for ² in order to achieve
ological research [6, 54]. This so-called learning rule
says that the synapses w~rl should be changed more if
convergent adaptation process [54]: vector quantization. The best way to find the mini-
Z t mum of E is gradient descent, i.e. movement in the
lim ²(t) = 0 and lim ²(t0 )dt0 = ∞. (9) vector-space along a line of maximal reduction of E
t→∞ t→∞ 0
according to change of prototypes w~ ~r :
Learning time t flows in parallel with increasing in- ² ∂E
dex k which belongs to sequentially arriving learn- w
~ ~r (t + 1) = w
~ ~r (t) − =
2 ∂w
~ ~r
ing examples ~xk . The second condition demands p
X
that decreasing of ² towards zero in time t (as =w
~ ~r (t) + ² (~xk − w
~ ~r (t))P (~xk ). (11)
prescribed by the first condition) must not be too k=1
fast. At the beginning of the learning process the This equation, obtained by inserting (10), is used
synapses may be empty or with random transmis- for every prototype w ~ ~r .
sion rates w~rl . The iteration process (6)/(7) using We can simplify equation (11) if the probability
function (8) is actually an alternative description of distribution P (~xk ) is not known explicitly what is
the signal-exchange determined by (4). often the case:
Two mathematical approaches to description of Ko- Here we have focused our attention only to the ad-
honen’s dynamics will be presented now. In most justment of the prototype w~ ~r0 which matches mostly
images, adjacent pixels are significantly correlated. with the current input-pattern ~xk . Input-patterns
Also if we have to store many similar input-patterns are presented successively during learning and in
(e.g., many faces) it is useful to store all correlated each step (12) is used.
pixels close to each other, to store all equal data only Equation (12) is equivalent to equation (7). In
once and to ignore all redundant data. order to get this generalization we would only have
Self-organizing mapping helps to solve this task. to add the neighborhood function h~r~r0 . In the most
It executes data storage by nearby-coding of similar general case which includes the neighborhood func-
information and at the same time it executes adap- tion and probability distribution we can write the
tive data compression. In self-organized maps data error function as a functional
X p
X
are stored in a more compact form, preserving all
E0 = h~r~r0 k~xk − w
~ ~r k2 P (~xk ). (13)
the most important information, so that the recalled
~
r~r0 k=1
data include as little error as possible comparing to
the original information. Usually we introduce a Dimension of vectors ~xk and w ~ ~r is n. k is the or-
measure of recall-error which should be minimized dinary number of an input-pattern which is to be
in order to achieve maximal data compression. learned or classified, respectively.
A suitable measure is the sum of all differences It is a special characteristics of Kohonen maps
between the input-patterns ~xk and the target- or that it is, say, three times more dense map (three
prototype-patterns w ~ ~r . By introducing a probabil- times more neurons) formed where there is three
ity density [27] of input-patterns P (~xk ) we can write times higher input-probability-distribution P (~xk ).
our measure as an expectation value E of the square So, we get a better map-resolution corresponding
error: to those receptive fields where there are more stim-
p
uli.
X
E(w)
~ = k~xk − w
~ ~r k2 P (~xk ). (10)
k=1 2.5 Principal component analysis
We want to minimize this expression in order to A very important way for feature extraction is the
compress data. This procedure is called optimal principal component analysis of the input-data [82,
special method of feature extraction and data com- organize themselves into functional and/or virtual
pression. hierarchies. Processes such as pattern recognition
A special case of mapping is search for an one- and storage, associations, adaptation to novel stim-
dimensional map. This mapping brings approxi- uli, extraction of the most relevant informational
mate solution to the so-called traveling salesman content, categorization, generalization, etc., can be
problem – how to connect many points with a short- modeled successfully using Hopfield neural networks
est trajectory [82, 30]. Kohonen solution of this (and their generalizations) which were presented in
classic optimization task is a Peano curve, a zig-zag the first part.
curve which homogeneously covers almost the whole In the second part Kohonen-type feed-forward
receptive field, and is sometimes fractal. DNA could neural network topologies were discussed which re-
be modeled as an one-dimensional map, and its alize these functions in another way. Symmetric as-
global encoding and decoding1 processes could be ap- sociative neural nets in the tertiary associative areas
proximated by the Kohonen self-organization. of the neocortex could not realize efficient higher-
All discussed approaches are used for modeling order associative and cognitive information process-
perception and motor action in living organisms. ing without feature extraction in the primary corti-
Connection vectors and cardinal neurons are to a cal areas and feature integration in the secondary
very large extent not genetically prespecified [31]. cortical areas [60]. The preprocessing of input data,
They evolve (like the DNA itself also, but much which is highly non-linear, is realized by Kohonen-
more gradually) under the selective influence from type networks of functional hierarchy.
environment. In a Kohonen neural network neurons interact
Present model is, of course, very important under the constraint of lateral inhibition in order
in neurophysics, but one can expect that it will to achieve specific responses of individual, localized
have significant impact on physics in general. It neurons to specific stimuli. Such specialization of
is namely an algorithm for dimension reduction neurons emerges as a consequence of underlying
in a complex system, and explains how from self- inter-neuronal interaction dynamics from ’compe-
organizing parallel-distributed processes localized tition for an information-representing role’. ’Win-
structures arise. ners’ in this ’competition’ are called cardinal neu-
rons. Thus, the Kohonen neural network provides
a mode of perceptual preprocessing as well as trig-
gering motor action [82, 70, 48]. For association ar-
3 Higher brain functions and con- eas of the neocortex, our Hopfield-type neural net-
sciousness: Synergetic neural works provide additional structure, accompanied by
networks and neuro-quantum virtual attractor-hierarchies, which is more appro-
priate for higher mental processing once the initial
coherence
stimulus analysis, primitive categorization and ab-
3.1 Synergetic conglomerates of straction processes have been realized and carried
patterns-qua-attractors: Simul- out by suitable Kohonen nets.
taneous operation at all levels In this part we will extrapolate presented theo-
ries, which were to a large extend tested also by
In the first part we have presented the most impor- author’s computer simulations [73] and were com-
tant aspects of cognitive processes at a micro-level pared with experimental (physiological) data, in or-
using symmetric associative neural networks, which der to model higher-brain functions and processual
1
During embryonal development DNA’s genetic informa-
basis of consciousness. Various associative, intuitive
tion is decoded in the process of continual external and inter- and even semantic processes could be approximated
nal environmental influences.
P
with models presented here; higher symbolic, lin- ck = N i i
i=1 vk q = h~vk , ~qi.
guistic, syntactic, logical, inferential processes, on (i is here the index of vector’s components and k
the other hand, could not be modeled using these the pattern index.) The order parameter ck acts as
P
models alone, but should be connected with sym- a coefficient in the series ~q = Pk=1 ck~vk .
bolic models of classical artificial intelligence [79] in In a special case of localized feature representation
order to make a hybrid model. Semantic and con- (as presented in the second part) the virtual order
ceptual networks, language structures, and speech parameter ck quantitatively represents the strength
production and recognition will not be presented of a cardinal neuron (a “dictator”-neuron) and its
here. corresponding pattern of neuronal population.
There are theoretical and experimental indica- Such a synergetic network is probably the most
tions (the last ones may perhaps be, for some sci- plausible model available for large-scale modeling
entists, still controversial) that consciousness is re- of (neo)cortical functions [75]. The neocortex has
lated with quantum systems [23, 14, 76]. Such ideas physiologically relatively uniform structure [28] and
were theoretically introduced by Wigner, von Neu- his enormous cognitive capabilities arise from col-
mann, Heisenberg, and later developed by Bohm, lective virtual processes which are still beyond the
Hiley, Stapp [91], Josephson, Penrose [71] and oth- methods of cognitive neuroscience [32, 19].
ers. The reasons for such a hypothesis, which In modeling higher brain functions, associative
receives increasing support in recent years (Tuc- and synergetic neural networks may be used in a
son multi-disciplinary conferences [42]), will be dis- generalized way, i.e., so that a generalized interpre-
cussed. tation of neurons is given. Generalized “neurons”
We feel that it is convenient to approach to the may be cortical columns or mini-columns, patterns-
research of quantum information processing with attractors of various orders (n)~v (nth virtual level),
a method of comparison of quantum and neural- etc. Generalized “connections” may be large-scale
network mathematical formalisms. Information physiological or virtual connections between cortical
processing capabilities of the neural network mod- areas, etc. [20].
els were intensively and successfully tested in com- As shown in figure 12, patterns-qua-attractors
puter simulations. Here their mathematical formal- are organized into a virtual hierarchy. Their dou-
ism will be compared to the quantum formalism in ble nature is shown: they are patterns (figure 12:
order to find similar, but (hypothetically) addition- left “boxes”) as well as attractors (figure 12: right
ally conscious, information processing capabilities “boxes”). Higher-order patterns contain lower-
of quantum systems. order patterns. A large hierarchy of patterns is
Haken’s synergetic model unites associative neu- a very flexible, mobile as well as sensitive and
ral nets and multi-layer neural nets. A multi- continually-changing structure. It consists of a set
level (originally three-level) synergetic network with of levels in two senses: firstly in the sense of vari-
intra-layer and inter-layer connections is established ous functional levels (as in feed-forward nets), and
[40, 39, 38]. Each layer is actually its own asso- secondly in the sense of various virtual levels of ab-
ciative network which can have functional (cardinal straction, or meaning, generality, and composition
cells in the second layer) or virtual interpretation [75]. We usually refer to these virtual levels of struc-
(order parameters ck – cardinal domains or neuronal ture in the following terms:
oligarchies in the second layer, or patterns ~vk in the – neurons (first biological level)
third layer). In Haken’s network order parameters – patterns (second level – first virtual level, gener-
ck measure the rate of overlap of a pattern ~vk with alized neurons)
the actual network-state ~q. So, ck is the projection – higher-order patterns, schemes, categories, col-
of ~vk onto ~q: lections of patterns, groupings of patterns; meta-
or at least constant phase difference. Such ordered cesses [75], seem to be too mechanistic, too dis-
collective oscillating behaviour is called coherence. crete and too deterministic for consciousness and
If patterns of neuronal oscillations are frequency- phenomenal qualia, i.e., qualitative perceptual ex-
locked and phase-locked, then they have a common periences.
feature, or something in common in more abstract On the other hand, all thought-processes,
sense, and represent an informational unity – a cat- including consciousness, seem to arise from
egory of higher order. complex-system-dynamics. Objects of conscious-
ness and stream of conscious thought seem to be
3.2 Synergetic levels in brain: neural, represented in some physical or at least informa-
sub-cellular, quantum and virtual tional (virtual) ”medium”. That ”medium” has to
processes be a complex-system which only is enough flexible,
fuzzy, adaptive, and has good self-organizing and
In a neural network neurons are constantly inter- recursive abilities. Because the mathematical
acting, and whole neuronal patterns are interacting. formalism of the neural network theory is confined
Neuron’s activities reflect each other. But it seems to the collective system-dynamics, it remains to
that also a very large and complex neural network a large extend valid also for complex systems of
with its recursive dynamics is not enough for the other basic elements [74]. So, our formal neurons
”real” consciousness as a global and unified self- and formal synaptic connections are not necessary
reflective process [42]. Neuronal patterns merely biological neurons and synapses (or, in models,
represent the objects of consciousness and take part artificial neurons—processors). There are various
in information processing concerning these objects, synergetic biological systems which may be mod-
but these patterns do not represent consciousness it- elled in a neural-network-like way and have relevant
self. It is a question how many different perceptions roles in micro-cognition processes on various levels:
are bound into an unified multi-modal experience – dendritic trees where neuronal dendrites have
where there are no traces of neuronal signaling and similar summation-tasks to that of a neuron
other material processes any more. Only a clear, [62, 39](Pribram);
relatively static qualitative experience remains. – subcellular structures: cytoskeleton, especially
There are indications that consciousness is con- microtubules, and other biomolecular systems of
nected with quantum phenomena [71, 91]. The electric or magnetic dipoles which may have a role
main reasons for this hypothesis are the following: of an interface between the neural and the quantum
– Quantum systems are the microscopic basis of level [41, 71];
all physical processes, including biological or psy- – webs of quantum particles (e.g., electrons) with
chophysical processes. All the classical world arises their spins [92, 93];
from the overall quantum background. – a ”continuum” of sub-quantum ”beables” (Bell)
– Quantum systems transcend even the division or ”hidden variables” [16, 44]; etc.
of particles and waves, or interactions, or fields
[23, 17]. Quantum systems, especially sub-quantum
systems, are holistic in nature [16] – they cannot
3.3 Neuro-quantum coherence needed
be satisfactorily analyzed into interacting funda-
for consciousness
mental elements, but act synthetically as indivisi-
ble parallel-distributed processes. As such, they are Consciousness itself (”pure” consciousness without
good candidates for correlates of the unity of con- an object of consciousness) may be associated with
sciousness. the quantum field, or better, with the ”overall
– Neural networks with their rigid neurons and sub-quantum sea” [33]. On the other hand, con-
synapses, in spite of their very subtle virtual pro- sciousness having some object as its content can-
not be associated merely with a specific quantum- ment, or our sensory apparatus) with the quantum
informational state. If a specific mental represen- system, the wave-function ”collapses” and a specific
tation is processed under control of consciousness, quantum eigenstate (a ”quantum pattern”) occurs
this specific representation, which is associated with as an attractor. So, there are quantum virtual struc-
a pattern of neural activity, is coupled or correlated tures also, and they cannot be reduced to a quan-
with a specific quantum eigen-wave-function. This tum eigenstate alone, because they occur only as a
quantum eigenstate was made explicit by the ”wave- result of interaction with a classical system. Thus
function collapse” which was triggered by corre- quantum virtual structures are (re)constructed as
sponding neural processes. a result of so-called quantum measurement where
The ”wave-function collapse” is a transition of the ”measurement apparatus” may be our sensory
the quantum state from a state described by a lin- and associative neural system directly, or a machine
ear combination of many quantum eigenstates to which is observed by that neural system. In both
a ”pure” state which is one eigenstate only. So, a alternatives the ”wave-function collapse” occurs as
superposition of many ”quantum patterns” is trans- a result of a specific interaction with a classical
formed into a single ”quantum pattern” only. system. The probability of the ”collapse” is very
Collapse of the wave-function means a selective much higher if the interaction is knowledge-based
projection from subconscious memory to the con- [74]. That’s like in the case of a radio: if we know
scious representation which was explicated from the the right frequency, we are able to receive the asso-
memory. There are two possible versions of mem- ciated information.
ory and memory-recall: the quantum one (just men-
tioned), or the classical neural one. In the first case
memory may be a parallel-distributed pattern in the 3.4 Comparison of the mathematical for-
system of synaptic connections. In the second case malism of associative neural network
memory could also be realized as a quantum holo- theory and quantum theory
gram, as a “pattern” in the implicate order of Bohm Main functional analogies and their informa-
and Hiley [17], or even as a set of ”parallel worlds” tional significance
according to the many-world interpretation of quan-
tum theory by Everett [26]. New consciousness studies give us reason to in-
Brain-mind is necessarily a multi-level phe- vestigate parallels between quantum processes
nomenon, because we cannot totally divide pure and neural-network-processes. Many mathemat-
consciousness from the object of consciousness ical analogies of the theory of associative neural
which may be an internal virtual image or a real ex- networks and the quantum theory can be found.
ternal object. We can then constitute the following Because we know that neural network simulations
overall scheme: Pure consciousness is of quantum perform well, we can search for similar effective
nature; virtual representations are associated with cognitive-like information-processing capabilities
neuronal patterns; external objects are of classical on the quantum level also. Let us make an overview
nature. Making a synthesis of these levels, neuro- of neuro-quantum analogies.
quantum coherence is constituted as a basis for
consciousness-about-some-object-of-consciousness. 1. NEURONAL-STATE-VECTOR ←→ QUAN-
Quantum mechanics governed by the Schrödinger TUM WAVE-FUNCTION:
equation doesn’t make attractors manifest until the In neural network theory the state of the system
”wave-function collapse” occurs. In that case, be- of neurons is described by q(~r, t) which denotes
cause of the interaction of a classical macroscopic the activity of an individual neuron (located at
system (either measurement apparatus, or environ- ~r) at time t. Neuronal patterns ~v are special
neuronal configurations ~q which represent some
p
X
where J(~r1 , t1 , ~r2 , t2 ) is the strength of an individ-
or J(~r1 , ~r2 ) = vk (~r1 ) vk (~r2 ). (23)
ual synaptic connection. J’s are the transmissions k=1
of real synaptic connections between two neurons
(spatially separated, but realized at the same time) Similarly, the Green function [13] is given as a sum
or correlations between states of different neurons of auto-correlations of individual quantum eigen-
at different times, represented at least virtually. states ψk :1
In non-relativistic quantum mechanics the dy- p
X
namical equation is the Schrödinger equation. Here G(~r1 , t1 , ~r2 , t2 ) = ψk (~r1 , t1 )∗ ψk (~r2 , t2 )
it is written, analogously to the neural equation k=1
p
X
(21), in Feynman’s form [44]:
or G(~r1 , ~r2 ) = ψk (~r1 )∗ ψk (~r2 ). (24)
Z Z k=1
Ψ(~r2 , t2 ) = G(~r1 , t1 , ~r2 , t2 )Ψ(~r1 , t1 )d~r1 dt1 . (22)
7. ADAPTIVE MEMORY IN CROSS-
CORRELATIONS: SYNAPTIC STRENGTHS
G(~r1 , t1 , ~r2 , t2 ) constitutes the Green function OBEYING THE GENERALIZED HEBB RULE
or an element of the quantum propagator [13]. OR DELTA RULE ←→ DENSITY MATRICES:
The propagator G is a matrix which describes a The kernels of dynamic equations (21) and (22) can
parallel-distributed transformation of the whole be written also as sums of cross-correlation func-
system from an initial state Ψ(~r1 , t1 ) to the final tions. In special cases they can be written as cross-
state Ψ(~r2 , t2 ). It is essentially a sort of Huygens’ correlation functions of a pattern and a difference
principle. The system transforms itself into a new between two patterns. In neural network theory we
state by exhibiting numerous internal interactions usually use cross-correlation of a novel pattern ~vd
between its constitutive ”quantum points” (some with a difference or discrepancy of this novel pat-
mathematical ”basic elements” of the system). tern ~vd and a desired or average pattern ~vl (or ~v̄)
Informationally, transformation (21) or (22) is where ~vl (or ~v̄) represents old knowledge [54, 56].
association. One pattern (concept, idea) is trans- A Hebb-like learning rule which is based on cross-
formed into another pattern (concept, idea) by a correlation of a (novel) pattern ~vd and a difference
collective informational interaction of input data ~vd − ~vl is called delta learning rule. It is used for
(encoded in the state vector, i.e. Ψ) with memory novelty filtering: new information, not represented
data (encoded in the system of connections or in the old neural pattern ~vl , is extracted and stored
interactions described by the Green matrix G). newly. Therefore the network finally adapts, i.e.,
optimally condenses and combines components of
6. MEMORY IN AUTO-CORRELATIONS OF new information (~vd −~vl ) with the whole knowledge
PATTERNS: SYNAPTIC STRENGTHS OBEY- collected up to now. The generalized Hebb learning
ING THE HEBB RULE ←→ GREEN FUNC- rule using cross-correlations of patterns is given as
TIONS: follows.
The kernels of dynamic equations (21) and (22) are p X
p
X
given as a sum of auto-correlation functions. The J(~r1 , ~r2 , t) = λkh vk (~r1 , t) vh (~r2 , t)
transmission-rate of an individual synaptic connec- k=1 h=1
X p Xp
tion J is determined by the Hebb learning rule as
a sum of auto-correlations between its two neurons or J(~r1 , ~r2 ) = λkh vk (~r1 ) vh (~r2 ). (25)
k=1 h=1
participating in various patterns ~vk :
1
Many physicists use another
√ definition of the Green func-
p
X tion, and there is an i (i = −1) in front of integrals of
J(~r1 , t1 , ~r2 , t2 ) = vk (~r1 , t1 ) vk (~r2 , t2 ) equation (6) in their case. The relation is: Gtheir−def =
i Gour−def .
k=1
This learning rule works also if ~vh is a difference the so-called diagonal parts of the density matrix
P
pattern ~vd −~vl or novelty pattern ~vd −~v̄ (delta rule), ρ(~r, ~r) = pk=1 ρkk | ψk (~r) |2 give a superposition of
or even if also ~vk is a difference pattern ~vg − ~vl or real-valued and thus measurable quantum probabil-
novelty pattern ~vg − ~v̄ (a case similar to equation ity densities with “weights” ρkk . ρkh is analogous to
(15) from section 2.5 where ~vd = ~vg ). λkh which acts as the so-called learning rate λr (see
Let us list presented versions of Hebbian cor- equation (16) and [40]).
relations once again. In the previous case (no. Interaction scheme no. 6 realizes auto-
6) strengths of individual synaptic connections association: completion of a neural pattern ~vk
were determined by a sum of auto-correlations or quantum pattern ψk from partial information.
(Hebb rule) of patterns, i.e., activity-correlations On the other hand, interaction web no. 7 realizes
of connected neurons which have a role in the same hetero-association: association form one pattern ~vk
patterns. In this case (no. 7) synaptic strengths are (or ψk ) to another pattern ~vh (or ψh ). Similar quan-
given by a sum of cross-correlations (generalized tum information processing using ρ-decomposition
Hebb rule) of patterns, i.e., activity-correlations was discussed by Ohya and Matsuoka [69, 65].
of neurons which have a role in different patterns. They related it to the Ising model.
If one of these patterns is a difference pattern
(difference between the first pattern and a reference 8. NEURONAL-PATTERN-RECONSTRUC-
or average pattern) then this generalized Hebb TION ←→ ”WAVE-FUNCTION COLLAPSE”:
rule is called delta rule. So, in the first case (no. The most important neuro-quantum analogy is the
6) a synapse represents correlation of neurons following. Pattern-reconstruction in a neural net-
constituting the same pattern, therefore we talk work
about pattern-auto-correlation, i.e., correlation p
X
of parts of the same pattern. In the second case q(~r, t) = ck (t) vk (~r )
(no. 7) a synapse represents correlation of neurons k=1
constituting two different patterns, therefore we =⇒ q(~r, t0 ) = vk0 (~r ) (27)
speak about cross-correlation or about correlation
between parts of two different patterns. is very similar to the ”collapse of the wave-function”
in a quantum system
Similarly to the present case (no. 7) of neural p
X
cross-correlation in equation (25), the density ma- Ψ(~r, t) = Ck (t) ψk (~r )
trix ρ is given as a sum of cross-correlations of indi- k=1
vidual quantum eigenstates ψk and ψh : =⇒ Ψ(~r, t0 ) = ψk0 (~r ). (28)
p X
X p For reasons presented in section 3.3 this is a very
ρ(~r1 , ~r2 ) = ρkh ψk (~r1 )∗ ψh (~r2 ). (26) important feature of cognitive processes. Processes
k=1 h=1
(27) and (28) are both a result of the influence from
Comparing equations (25) and (26), and knowing tors or Green functions do. They are active-memory descrip-
that the generalized Hebb rule for J realizes effi- tors, but the density matrix is a “passive” description (like an
cient associative memory, we may infer about in- “archive”) of statistical and correlational properties of quan-
tum states in the present and history of the system. However,
formation processing capabilities of ρ – the quan- the density matrix may represent an important quantitative
tum probability density matrix or quantum statis- description of relations of eigenstates (patterns) which are
tical operator, respectively.1 Just note also that crucial if a quantum system is harnessed for information pro-
cessing, or here specifically for cross-correlations or hetero-
1
Of course, the density matrix itself does not describe associations. Authors like Alicki, Ohya and Matsuoka have
(or even execute, in simulations) the information-processing theoretically described quantum information transmission us-
input-to-output transformation in such a way as propaga- ing density matrices and their tensor products.
the system’s environment. The environment selects of neurons ~q. The so-called impulse (p-) representa-
those neural/quantum pattern which is the most tion can, on the other hand, be associated with the
similar (or is correlated) to the state of environ- system of interactions J which regulates all transfor-
ment. mations of the network-state. Thus, the quantum
Neural-pattern-reconstruction and wave-function duality (particles/waves, or particles/interactions,
collapse are results of a transition from the im- or fermions/bosons, respectively), has a counterpart
plicate order (Bohm’s term for latent, implicit, in neural duality (neurons/synaptic connections, or
inactive, potential non-Shannon information only) neurons/interactions).
to the explicate order (manifest, active, realized Sometimes we find that a network where neurons
Shannon information) [17]. The implicate order and synapses are exchanged has a similar behaviour
represents a combination of very many possible as the original one. Thus, in general there is a
states or processes (left-hand side equation in (27) symmetry of neurons and synapses (or interactions)
and (28)). It is analogous to a set of so-called which represents invariance under neurons-synapses
”parallel worlds” or parallel sub-branches of the exchange. If we compare, for example, simulations
general wave-function proposed by Everett [26]. of associative neural networks by this author [73]
Explicate order, on the other hand, represents a and simulations by Menneer [66]1 , we see that ex-
state or process which is at a moment physically plicit processing, which is realized in the system of
actualized (right-hand side equation in (27) and neurons in this-author’s case, is similar to those ex-
(28)). It is ”chosen” from a set of potential plicit dynamics which is realized in the system of
(implicate) states, or is a result of their optimal connections in some Menneer’s cases. Menneer in
”compromise”. “Taking choice” may be an “act some cases performs “collapse” (pattern selection
of consciousness” which takes some information and reconstruction) in the systems of connections,
out from memory or subconsciousness (implicate but this author performs it in the system of neu-
order), and makes it explicit and conscious. rons. She also emphasizes analogy of this collective
dynamics with processes during quantum measure-
9. NEURAL DUALITY AND QUANTUM DU- ment. We cannot extract information in the system
ALITY: of neurons and in the systems of synaptic connec-
An interesting neural analogy of the Heisenberg un- tions at the same time – not in both systems. This
certainty principle of quantum mechanics is repre- is a neural version of uncertainty case. We have to
sented by an inability of simultaneous determination make a “collapse” with output information repre-
of patterns in the system of neurons ~q and of pat- sented either in the system of neurons or in the sys-
terns in the system of interactions or connections J. tem of synapses. Without a “collapse” the interplay
We are unable to make a pattern explicit in a sys- of neurons and connections is always an indivisible
tem of neurons, and to make a pattern explicit in holistic process.
the system of connections at the same time. Only Like in classical dynamics, for integrable sys-
one pattern, which is at the moment realized in the tems, there is a complete equivalence between the
system of neurons, is made explicit. All the others Schrödinger equation for Ψ and the Liouville-von
are present only implicitly in the system of interac- Neumann equation for ρ [81]. Both kinds of
tions, i.e., in the coupling-rate. In cognition we have description are complementary on quantum as well
a similar situation: we can be aware of one pattern as classical (e.g., neural) level.
only which has been extracted from memory. Other
memory patterns remain unconscious and implicit.
The so-called position (x-) representation of 1
Comparing with [69, 65] similar symmetry-observations
quantum theory can be approximated by the system can be made concerning the density matrix ρ and its decom-
positions.
10. QUANTUM AND NEURAL WAVELET analogously to the neural-net-state expansion (17)
UNCERTAINTY PRINCIPLES: with coefficients (19). Gabor wavelets are, however,
Some more formalized uncertainty analogies can in general not orthogonal.
also be found. In the neural network and early vi- Daugman [22] presented psycho-physical ana-
sion theories we have uncertainty principles which logues of Gabor uncertainty principle for 2 di-
1 1
are similar to the Heisenberg uncertainty princi- mensions: ∆x∆ux ≥ 4π and ∆y∆uy ≥ 4π ,
1
ple. Gabor proofed [62] that the Heisenberg un- which give together ∆x∆y∆ux ∆uy ≥ 16π2 .
certainty principle is a special case of his general MacLennan [62] naturally added Gabor uncer-
1
frequency/time uncertainty principle: ∆f ∆t ≥ 1. tainty principle for 3 dimensions: ∆x∆ux ≥ 4π ,
1 1
Error in measuring frequency ∆f is inverse propor- ∆y∆uy ≥ 4π , ∆t∆f ≥ 4π , which give to-
1
tional to the error in measuring time ∆t. Multi- gether ∆x∆y∆t∆ux ∆uy ∆f ≥ 64π 3 . So, in the 3-
plying both sides of Gabor’s inequality by Planck dimensional case each information cell, represent-
constant h, and knowing E = hf , we obtain the ing a “quantum of information”, is a 6-dimensional
well-known Heisenberg’s version: ∆E∆t ≥ h. Ga- hypercube (∆x, ∆y, ∆t, ∆ux , ∆uy , ∆f ) localized to
bor uncertainty principle ∆f ∆t ≥ 1 is applicable spatial area ∆x∆y and to temporal interval ∆t,
whenever simultaneous measurements of an observ- and furthermore tuned to spatial frequency bands of
able and of its Fourier transform are made [62]. Pre- width ∆ux and ∆uy , and to (temporal) frequency
viously mentioned x/px and E/t Fourier or uncer- band of width ∆f .
tainty pairs of conjugate variables are thus similar This overview of uncertainty relations of the
to the f /t pair of Gabor’s information theory and theory of processing in the visual cortex showed
x/ux pair of Daugman’s vision theory. In Daug- that the uncertainty relations in general, including
man’s theory [22] spatial frequency ux determines Heisenberg’s quantum-mechanical version, are con-
the wavelet’s location in frequency domain so as nected with informatics of parallel-distributed com-
x determines location in ordinary spatial domain. plex systems in general, even classical ones such as
It is physiologically supported that Gabor’s and associative neural nets. Uncertainty relations arise
Daugman’s theories are applicable for representing as a result of observational perturbation of quantum
spatio-temporal visual images and information fil- systems, but it may emerge also in the case of clas-
tering in the visual cortex [22]. This is a large-scale sical (neural) complex systems. This observational
field-theory-like approach in contrast to usual dis- perturbation has much more evident and exact con-
crete neural-net-models. sequence (such as conjugate pairs of non-commuting
The Gabor inequality changes to equality observable operators) in quantum complex systems,
∆f ∆t = 1 (minimal uncertainty) if neural wavelets because of their microscopic scale, than in classical
are modeled by complex oscillatory functions having complex systems.
Gaussian envelope. These are Gabor functions ψjk Bohm noted that if we are starting to think about
which constitute basic information cells represented our own thoughts, we change them with this self-
in space and spatial frequency areas observation act. That’s uncertainty of conscious
à ! self-reflection which Bohm has related to quantum
π(t − j∆t)2 uncertainty [15].
ψjk = exp − exp (2πik∆f (t − j∆t)) .
α2
(29)
Discussion of discrepancies in presented
They are similar to quantum coherent states or wave
analogies
packets. An arbitrary function Ψ can be represented
by elementary Gabor functions ψjk as their linear There is a difference between equations (23) and
P P∞
superposition Ψ = ∞ j=−∞ k=−∞ Cjk ψjk , where (24): the latter one includes complex-valued vari-
Cjk = hΨ, ψjk i/ k ψjk k (if wavelets are orthogonal), ables which are necessary for the quantum case, but
just possible for the classical neural-net case. That and mathematical analogies between quantum and
prevents us from claiming that quantum systems are neural levels in the case of oscillatory dynamics of
a very complex sort of ”neural networks”. Quantum neurons. Specifically, if we have an associative net-
systems are much more subtle and flexible, because work of coupled neurons with oscillatory activities
they don’t have well-defined fundamental particles [40, 97, 63] complex-valued mathematical formal-
and interactions. So, no well defined basic formal ism becomes relevant for neural nets also. Our pre-
“neurons” or formal “connections” can be found – sented neuro-quantum analogies, which in this case
no well-defined “hidden variables” of Bohm [16]. contain information about phase-correlations, then
Anyway, in spite of this difficulty, the mathematics become much more important. In fact, John G.
of the usual Copenhagen interpretation of quantum Sutherland has created a simulation of such a net-
mechanics is essentially very close to the mathemat- work of neurons with oscillatory activities, called
ics of associative neural networks. So, the limits of “Holographic Neural Technology”, which processes
our or similar useful [7] analogies are also limits of very successfully. A condensed presentation by
the orthodox (Copenhagen) quantum theory. Sutherland, entitled “Holographic model of mem-
Replacement of complex-valued quantum equa- ory, learning and expression”, was published in 1990
tions with a system of real-valued equations [15, 17] in the International Journal of Neural Systems, vol.
or introduction of complex numbers into the neu- 1, pages 256-267, which is highly recommended.
ral network theory (e.g., networks of coupled oscil-
lators [40, 63, 97, 90, 8], quantum neuro-holography Some further analogies
[85, 84, 86]) would support the search for further
11. QUANTUM NEURAL HOLOGRAPHY:
analogies with the neural network theory. Complex-
WAVELET-MIXING OR CONVOLUTION:
number variables in quantum formalism obviously
In neural networks memory consists of
have a very deep meaning. Ψ∗ as opposed to Ψ
(auto)correlations between patterns. In quan-
is connected with the phase-conjugated or time-
tum mechanics phase differences between different
reversed dynamics. So, in quantum systems we
parts of the wave-function may have a similar role
have a superposition of forward and backward
[15]. They control the time evolution of probability
time-evolution. In neural networks this time-flow-
distribution involving interference of the contribu-
superposition is transformed into superposition of
tions of different stationary eigen-wave-functions.
associative transformations in the opposite direc-
In quantum neuro-holography we have interference
tions in space, or the superposition is completely
patterns of coherent neural wavelets. Here phase
collapsed into single directed dynamics (time se-
shift ϕ encodes the degree of correlation between
quence of patterns in the direction of the arrow of
components of coherent wavelets. On the other
time). Observed quantum systems reflect an intrin-
hand, their amplitude encodes the significance
sic complementary duality which is evident also in
of synchronized wavelet components [85]. In-
the wave/particle double nature or in location and 0 0
deed, wavelet mixing ψk (t)dt ⊗ ψh (t )dt , (mapped
momentum representations. Probably this appar-
onto) spatio-temporal cross-correlation function
ent duality, and uncertainty principles also, are a R 2πiνϕt
e ψk (t + ∆t)ψh (t)dt1 [64], or convolution,
consequence of an active observer’s participation in
etc., are examples of bilinear forms similar to
the observed quantum process.
the generalized Hebb learning rule used in neural
In spite of the fact that quantum systems are de-
networks. So, quantum holography and theory
scribed by a dual complex-valued (real plus imagi-
of associative neural nets (in particular, if neu-
nary) description, but the usual neural net models
rons have oscillatory activities) have significant
only in a real-valued description so far, we must
mathematical analogies which are a consequence
emphasize that there are very interesting functional
1
ν denotes a fixed spectral parameter.
of similar collective dynamics in both complex as in conventional associative networks [40] and
systems. similar phase-coupling or frequency-coupling (co-
In generalized neural network theory we have herent behaviour) as with wavelets in holography.
“matrices of (synaptic) connections” which encode
correlations of two neural patterns. In holography, 13. “COLLAPSED” QUANTUM SUPERPOSI-
on the other hand, we have holograms which en- TION ←→ FORMAL NEURAL NETWORK:
code correlations of two quantum-optical signals, Quantum holography, however, introduces some-
i.e., their interference. In both theories we use sim- thing more. It gives opportunity for introduction
ilar mathematics for similar information processing of complex conjugate form of a wavelet. Complex
as long as we can neglect details in the structure of conjugate wavelet ψ ∗ is a time mirrored version of
basic elements of the network. the original wavelet ψ.
Elements of the system are, of course, different The well-known quantum bilinear form ψψ ∗ rep-
again. This brings differences between quantum resents a combination of a state vector (wave func-
holography and associative neural nets, but it tion) ψ with its phase-conjugated (time-mirrored)
seems that brain realizes quantum-holographic-like form ψ ∗ . Together they give real-valued probabil-
and neural-network-like processing on different ity density | ψ |2 = ρ, in the case of a pure state.
scales. So, in spite of differences in “anatomical” For “mixtures” a weighted sum of ψk ψk∗ has to be
details, we may speak about fractal-like multi-level written [81] (in general this gives expression (26)).
synergetic processing in brain. This is caused by Thus, in quantum world we have superpositions of
similar “sociology” in networks on various bio- many possible networks, processes and “arrows of
physical levels, i.e., quantum, sub-cellular, neural, time”.
virtual or conceptual networks or “holograms”. Only one network, process and “arrow of time”
is, however, realized explicitly during measurement
12. PHASE RELATIONS IN QUANTUM process, and this makes situation similar to the case
HOLOGRAMS ←→ STRENGTHS OF FORMAL of neural nets. To summarize roughly, a generalized
SYNAPSES: (mathematical) neural network emerges as a result
Changing the phase relations between eigen- of “collapse” of a quantum superposition of implicit
wave-functions (e.g., quantum-optical signals) “networks”. To say as simple and short as possible:
is analogous to the learning-process in neural mathematical neural network (network of mathe-
networks where new pattern correlations are added matical points or formal “neurons” which interact
to the synaptic correlation-matrix. As mentioned, intensively) is a “collapsed” quantum system.
conventional neural network theory does not incor- Biological neural network evolves as a macroscopic
porate neurons with oscillating activities, although replica of these fundamental dynamical principles.
this is observed in brain. Only some more advanced
neural theories introduce oscillating neuronal 14. Fourier-like CORRESPONDENCE RULES
activities and put relevance to phase-relations or between the above equations can be shown first for
phase-correlations between neurons with oscillating the neural case (30), and then for the quantum case
firing rates. Phase-correlations are an essential (31), using the following calculation.
ingredient of (quantum) holography, therefore asso-
ciative networks with oscillating neuronal activities q(~r2 , t2 )
are certainly more similar to holographic devices Z Z
than conventional associative neural networks. If = J(~r1 , t1 , ~r2 , t2 ) q(~r1 , t1 ) d~r1 dt1
we study networks with oscillating neuronal activi- Z Z X
p
ties, we find similar associative memory capabilities = [ vk (~r1 , t1 ) vk (~r2 , t2 )] q(~r1 , t1 ) d~r1 dt1
k=1
p
X Z Z p
X Z Z
= vk (~r2 , t2 )[ vk (~r1 , t1 ) q(~r1 , t1 ) d~r1 dt1 ] = λk vk (~r2 , t2 )[ vk (~r1 , t1 ) q(~r1 , t1 ) d~r1 dt1 ]
k=1 k=1
p
X p
X
=⇒ q(~r, t) = ck (t) vk (~r, t) =⇒ q̇(~r, t) = λk ck vk (~r, t)
k=1
k=1
Xp
Xp
or q(~r, t) = ck (t) vk (~r ) (30) or q̇(~r, t) = λk ck vk (~r ). (32)
k=1
k=1
In the last step we have used definition (19) for
In quantum case we can make the following calcu-
time-dependent patterns. For the stationary case
lation starting from the time-dependent Schrödin-
we delete the time integral.
ger equation with the Hamiltonian operator Ĥ. In
Ψ(~r2 , t2 ) the last step we use the stationary Schrödinger
Z Z equation Ĥ ψk (~r) = Ek ψk (~r) where Ek is the
= G(~r1 , t1 , ~r2 , t2 ) Ψ(~r1 , t1 ) d~r1 dt1 energy-eigenvalue.
Z Z X
p
ih̄Ψ̇(~r, t) = ĤΨ(~r, t)
= [ ψk (~r1 , t1 )∗ ψk (~r2 , t2 )] Ψ(~r1 , t1 ) d~r1 dt1 X
k=1
= Ĥ[ Ck ψk (~r )]
Z Z k
p
X X X
= ψk (~r2 , t2 )[ ψk (~r1 , t1 )∗ Ψ(~r1 , t1 ) d~r1 dt1 ] = Ck Ĥψk (~r ) = Ck Ek ψk (~r ). (33)
k=1 k k
p We thus obtain
X
=⇒ Ψ(~r, t) = Ck (t) ψk (~r, t) p
k=1
i X
Ψ̇(~r, t) = − Ek Ck ψk (~r )
Xp h̄ k=1
or Ψ(~r, t) = Ck (t) ψk (~r ) (31) p
k=1
i X
or Ψ̇(~r, t) = − Ek Ck Ψk (~r, t) (34)
h̄ k=1
15. The following version of dynamic equations
can be derived from neuro-synergetic equations (21) Here Ek (an eigenvalue which represents result of
and (23) which are here somewhat generalized. The a measurement of an observable quantity, i.e., en-
so-called attention parameter λk is added [40]. At- ergy) has a similar role as λk , the neuro-synergetic
tention parameter λk of the pattern ~vk is an eigen- attention parameter. Note the interdependence of
value of the matrix J with eigenvectors ~vk : J ~vk = the process of attention and observation which is
λk ~vk . (In quantum case, analogously, ψk are eigen- here, probably not completely by chance, replicated
states of the density operator ρ with eigenvalues in the mathematical formalism.
ρk [65]). We insert a generalized version of the
Hebb rule (23) into the time-dependent equation 16. The analogy no. 6 can be extended to rela-
(21). In the second step we use definition (19) for tivistic domain also. In the relativistic case, where
time-dependent patterns. For stationary case we ψk are 4-dimensional vectors, the role of the Green
can delete the time integral. function G is realized by the S − matrix (scattering
q̇(~r2 , t2 ) matrix) [13]. Its elements are:
Z Z X
p p X
X 2
= [ λk vk (~r1 , t1 ) vk (~r2 , t2 )] q(~r1 , t1 ) d~r1 dt1 S(~r1 , t1 , ~r2 , t2 ) = − ψkj (~r1 , t1 )∗ ψkj (~r2 , t2 )
k=1 k=1 j=1
equation (37), the following “non-local probability macroscopic environment in a specific and flexible
continuity equation” is obtained: manner by using the biological neural network as
µ ¶ a macro-micro-interface and a (subconscious) pre-
∂ρ ρ∇S
+ div processor for an unified conscious experience which
∂t m
involves neuro-quantum coherence.
µZ ¶
= 2Im V (~r1 , ~r2 , t)Ψ(~r1 , t)∗ Ψ(~r2 , t)d~r1 d~r2 . (39)
vanced system-based artificial intelligence in form [4] R. Amoroso. The Extracellular Containment of
of neural, quantum [59, 25, 21], and neuro-quantum Natural Intelligence. Informatica 19, 585-590
[49, 50, 51] or other hybrid parallel (synergetic) (1995).
computers. [5] R. Amoroso. Modeling the Heisenberg Ma-
Although there is no strict mathematical iso- trix: Quantum Coherence and Thought at the
Holoscape Manifold and Deeper Complementar-
morphism in neural and quantum formalisms, sec-
ity. In: Scale in Conscious Experience: Is the
tion 3.4 lists remarkable parallels which enable Brain too Important to Be Left to Biologists to
us to use neural-net processing as a metaphor Study? Edited by K. Pribram. (Lawrence Earl-
for quantum network processing. These neuro- baum, Hillsdale (NJ), 1995).
quantum analogies are an original contribution [6] Neurocomputing – Foundations of Research.
which enable development of quantum neural-net- Edited by J.A. Anderson, E. Rosenfeld. (MIT
like information-processing ”algorithms”. Indeed, Press, Cambridge (MA), 1988).
in [78] we presented our new quantum associative [7] J.B. Anderson. A random-walk simulation of the
network model.1 Schrödinger equation: H3+ . J. Chem. Phys. 63,
no. 4, 1499-1503 (1975).
[8] B. Baird. Bifurcation and category learning in
network models of oscillating cortex. Physica D
Acknowledgements 42, 365-384 (1990).
[9] P. Baldi, K. Hornik. Neural Networks and Prin-
TEMPUS grant (JEP-4310-93) for my research in cipal Component Analysis. Neural Netw. 2, 53
London, OSI grant, and discussions with Prof. K.H. (1989).
Pribram, Prof. B.J. Hiley, Prof. L.I. Gould, Prof. [10] W. Banzhaf, H. Haken. An energy function for
A.O. Župančič, Dr. J. Šlechta, Dr. T. Matsuoka specialization. Physica D 42, 257-264 (1990).
and T. Menneer are gratefully acknowledged. Many [11] H. Barlow. Intraneuronal information process-
thanks also to P. Ečimovič for his significant help ing, directional selectivity and memory for spatio-
in preparation of the first part, and to Prof. A. temporal sequences. Network: Comp. in Neural
Linkevich, Prof. A. Ezhov and Dr. V. Samuilov. Sys. 7, 251-259 (1996).
[12] R. Beale, T. Jackson. Neural Computing (An In-
troduction). (Adam Hilger, Bristol, 1990.)
References [13] J.D. Bjorken, S.D. Drell. Relativistic Quantum
Fields. Chapter 6: Propagator Theory. (McGraw-
[1] S.V.B. Aiyer, M. Niranjan, F. Fallside. Theoreti- Hill, New York, 1965).
cal investigation into the performance of the Hop- [14] G.J.F. Blommestijn. A Theory on the Relation
field model. IEEE Transact. on Neural Netw. 1, between Quantum Mechanical Reduction Process
204 (1990). and Consciousness. Presented at the 2nd Conf. on
[2] S.-I. Amari, K. Maginu. Statistical Neurodynam- Consciousness, Tucson (AZ), 1996 (preprint).
ics of Associative Memory. Neural Netw. 1, 63-73 [15] D. Bohm. Quantum Theory. (Constable and Co.,
(1988). London, 1954).
[3] D. Amit. Modeling Brain Functions (The world of [16] D. Bohm. Wholeness and Implicate Order. (Rout-
attractor neural nets). (Cambridge Univ. Press, ledge and Paul Kegan, London, 1980).
Cambridge, 1989). [17] D. Bohm, B.J. Hiley. The Undivided Universe
1
(An ontological interpretation of quantum the-
It triggered, since 1998, several sub-conferences on quan- ory). (Routledge, London, 1993).
tum neural-like nets as parts of multi-conferences in USA
(IJC Information Sciences ’98 [101], ’99, and NIPS 2000), [18] H. Bourlard, Y. Kamp. Auto-Association by Mul-
Russia (Neuroinformatika ’99, 2000) and Belgium (CASYS tilayer Perceptrons and Singular Value Decompo-
’99, 2000). sition. Biol. Cybern. 59, 291 (1988).
[19] F. Brentano. Psychology from an Empirical [35] L.I. Gould. Reflections on the Relevance of Non-
Standpoint. (Routledge and Kegan Paul, London, locality to Cognitive science and the Philosophy
1973). of Mind. In: New Directions in Cognitive Sci-
[20] Y. Burnod. An Adaptive Neural Network: the ence, Proc. Int. Conf. in Lapland. Edited by P.
Cerebral Cortex. (Prentice Hall, London, 1990). Pylkkänen, P. Pylkkö. (Finnish AI Soc., Helsinki,
1995, pp. 104-114).
[21] R.L. Chrisley. Quantum Learning. In: New Di-
rections in Cognitive Science, Proc. Int. Conf. [36] H. Gutfreund. From Statistical Mechanics to
in Lapland. Edited by P. Pylkkänen, P. Pylkkö. Neural Networks and Back. Physica A 163, 373-
(Finnish AI Soc., Helsinki, 1995, pp. 77-89). 385 (1990).
[22] J.G. Daugman. Uncertainty relation for resolu- [37] H. Haken. Advanced Synergetics (Instability Hier-
tion in space, spatial frequency, and orientation archies of Self-Organizing Systems and Devices).
optimized by 2-D visual cortical filters. J. Opt. (Springer, Berlin, 1987).
Soc. Amer. A 2, no. 7, 1160 (1985). [38] Neural and Synergetic Computers. Edited by H.
Haken. (Springer, Berlin, 1988).
[23] P.C.W. Davies, J.R. Brown. The Ghost in
the Atom. (Cambridge Univ. Press, Cambridge, [39] Synergetics of Cognition. Edited by H. Haken, M.
1986). Stadler. (Springer, Berlin, 1989).
[24] J.S. Denker. Neural network models of learning [40] H. Haken. Synergetic Computers and Cogni-
and adaptation. Physica D 22, 216 (1986). tion (A Top-Down Approach to Neural Nets).
(Springer, Berlin, 1991).
[25] D. Deutsch. Quantum computation. Phys. World,
pp. 57-61 (June 1992). [41] S.R. Hameroff. Quantum coherence in micro-
tubules: a neural basis for emergent conscious-
[26] The Many-Worlds Interpretation of Quantum
ness? J. Consciousn. Stud. 1, no. 1, 91-118
Mechanics. Edited by B.S. DeWitt, H. Graham.
(1994).
(Princeton Univ. Press, Princeton, 1973.)
[42] Towards a Science of Consciousness, Tucson I.
[27] Models of Neural Networks (Series ”Physics of
Edited by S.R. Hameroff, A.W. Kaszniak, A.C.
NN”). Edited by E. Domany, J.L. van Hemmen,
Scott. (MIT Press, Cambridge (MA), 1996.)
K. Schulten. (Springer, Berlin).
[43] R. Hecht-Nielsen. Neurocomputing. (Addison-
[28] M. Ebdon. Is the Cerebral Neocortex an Uniform
Wesley, Reading, 1990).
Cognitive Architecture? Mind & Language 8, 368
(1993). [44] B.J. Hiley, F.D. Peat. Quantum Implications (Es-
says in Honour of David Bohm). (Routledge,
[29] R. Feynman. The Theory of Positrons. Phys. Rev.
London, 1987).
76, no. 6, 749 (1949).
[45] J.J. Hopfield. Neural networks and physical sys-
[30] J.C. Fort. Solving a Combinatorial Problem via
tems with emergent collective computational
Self-Organizing Process: An Application of the
abilities. Proceed. Natl. Acad. Sci. USA 79, 2554-
Kohonen Algorithm to the Traveling Salesman
2558 (1982).
Problem. Biol. Cybern. 59, 33 (1988).
[46] J.J. Hopfield. Neurons with graded response have
[31] A. Gierer. Spatial Organization and Genetic In- collective computational properties like those of
formation in Brain Development. Biol. Cybern. two-state neurons. Proceed. Natl. Acad. Sci. USA
59, 13 (1988). 81, 3088 (1984).
[32] G. Globus: Towards a Noncomputational Cog- [47] M. Jibu, K. Yasue. Quantum Brain Dynamics
nitive Neuroscience. J. Cognit. Neurosci. 4, 299 and Consciousness. (John Benjamins, Amster-
(1992). dam / Philadelphia, 1995).
[33] A. Goswami. Consciousness in Quantum Physics [48] M. Kaipainen, P. Toiviainen, J. Louhivuori. A
and the Mind-Body Problem. J. Mind & Behav. Self-Organizing Map that Recognizes and Gen-
11, no. 1, 75 (1990). erates Melodies. In: New Directions in Cognitive
[34] L.I. Gould. Nonlocal Conserved Quantities, Bal- Science, Proc. Int. Conf. in Lapland. Edited by P.
ance Laws, and Equations of Motion. Int. J. Pylkkänen, P. Pylkkö. (Finnish AI Soc., Helsinki,
Theor. Phys. 28, no. 3, 335-364 (1989). 1995, pp. 286-315).
[49] S.C. Kak. Information, Physics, and Computa- [67] J.L. McClelland, R. Rumelhart, PDP group: Par-
tion. Foundat. Phys. 26, no. 1, 127-137 (1996). allel Distributed Processing – Explorations in the
[50] S.C. Kak. On quantum neural computing. In- Microstructure of Cognition: vols. 1, 2. (MIT
form. Sci. 83, 143-160 (1995). Press, Cambridge (MA), 1986).
[51] S.C. Kak. Quantum Neural Computing. Ad- [68] M.E. Nelson, J.M. Bower. Brain maps and par-
vances in Imaging & Electron Phys. 94, 259-313 allel computers. Trends in Neurosci. 13, no. 10,
(1995). 403 (1990).
[52] I. Kanter. Potts-glass models of neural networks. [69] M. Ohya. Fractal Dimensions of States. In: Quan-
Physical Rev. A 37, no. 7, 2739-2742 (1988). tum Probability and Related Topics IV. (World
Scientific, Singapore, 1991, pp. 359-369).
[53] T. Kohonen. Self-organized formation of topolog-
ically correct feature maps. Biol. Cybern. 43, 59 [70] R.G. Palmer, J. Hertz, A. Krogh. Introduction
(1982). to the Theory of Neural Computation. (Addison-
Wesley, Redwood City, 1991).
[54] T. Kohonen. Self-Organization and Associative
[71] R. Penrose. Shadows of the Mind (A Search for
Memory. (Springer, Berlin, 1984).
the Missing Science of Consciousness). (Oxford
[55] T. Kohonen. Adaptive, Associative, and Self- Univ. Press, Oxford, 1994).
Organizing Functions in Neural Computing.
[72] P. Peretto. An Introduction to the Modeling of
Appl. Optics 26, 23 (1987).
Neural Networks. (Cambridge Univ. Press, Cam-
[56] T. Kohonen. An Introduction to Neural Comput- bridge, 1992).
ing. Neural Netw. 1, 3 (1988).
[73] M. Peruš. All in One, One in All (Brain and
[57] A.D. Linkevich. Anticipation, perception, lan- Mind in Analysis and Synthesis). (DZS, Ljubl-
guage, mind and nonlinear dynamics. Informatica jana, 1995). M. Peruš. Bio-Networks, Cognition
21, 435-464 (1997). and Consciousness. (DZS & Satjam, Ljubljana
[58] R. Linsker. Self-Organization in a Perceptual & Maribor, 2000). (Both books are in Slovene.)
Network. Computer, pp. 105-117 (March 1988). [Quantitative details in: Neurophysical Founda-
[59] S. Lloyd. Quantum-Mechanical Computers. Sci. tions of Cognitive Science; unpubl. manuscript.]
Amer., pp. 44-50 (October 1995). [74] M. Peruš. Analogies of neural and quantum pro-
[60] A.R. Lurija. Foundations of Neuropsychology. cessing – Consequences for cognitive science.
(Russ. orig.: Osnovyi Neyropsihologii. Izdat. In: New Directions in Cognitive Science, Proc.
moskov. univ., Moscow, 1973). Int. Conf. Edited by P. Pylkkänen, P. Pylkkö.
(Finnish AI Soc., Helsinki, 1995, pp. 115-123).
[61] R.J. MacGregor. Neural and Brain Modeling.
(Academic Press, San Diego, 1987). [75] M. Peruš. Synergetic Approach to Cognition-
Modeling with Neural Networks. In: Bilder im
[62] B. MacLennan. Information Processing in the Geiste. Edited by K. Sachs-Hombach. (Rodopi,
Dendritic Net. Chapter 6 in: Rethinking Neural Amsterdam & Atlanta, 1995, pp. 183-194).
Networks (Quantum Fields and Biological Data.
Edited by K. Pribram. (Lawrence Erlbaum As- [76] M. Peruš. System-processual backgrounds of con-
soc., Hillsdale (NJ), 1993). sciousness. Informatica 21, 491-506 (1997). [Al-
ternatively, see in Slovene: Concepts of Holis-
[63] C.L.T. Mannion, J.G. Taylor. Information Pro- tic ”Models” of Consciousness (Bohm’s Quantum
cessing by Oscillating Neurons. In: Coupled Os- Implications, Holograms and Neural Networks).
cillating Neurons. Edited by J.G. Taylor, C.L.T. J. Critics of Sci. 174, 11-22 (1995).]
Mannion. (Springer, London, 1992).
[77] M. Peruš. Neuro-Quantum Parallelism in Mind-
[64] P. Marcer, W. Schempp. The brain as a conscious Brain and Computers. Informatica 20, 173-183
systems. J. General Sys. 27, 231-248 (1998). (1996).
[65] T. Matsuoka, M. Ohya. Fractal Dimensions of [78] M. Peruš. From neural to quantum associative
States and Application to Ising model. Reports networks: A new quantum ”algorithm”. AIP
Math. Phys. 36, no. 2/3, 365-379 (1995). Conf. Proceed. of CASYS’99 (Computing An-
[66] T. Menneer. Quantum-inspired two-layer neural ticipatory Systems – 3rd Int. Conf., Liege, Bel-
networks. Tech. rep., Univ. of Exeter, Comp. Sci. gium), vol. 517. Edited by D. Dubois. (Amer-
Dept., 1996 & Tucson II conf. ican Institute of Physics, Melville (NY), 2000,
pp. 289-295). [Alternatively: M. Peruš: Pro- [91] H.P. Stapp. Quantum Propensities and the
ceed. JCIS’98, vol. 2, RTP, NC, USA, pp. 197- Brain-Mind Connection. Foundat. Phys. 21, no.
200. Neural Netw. World 10 (no. 6), in press 12, 1451 (1991).
(2000). Appl. Math. Lett. 13, no. 8, 31-36 (2000) [92] A. Stern. Matrix Logic and Mind (A probe into
(with S.K. Dey).] the unified theory of mind and matter). (North-
[79] Foundations of Cognitive Science. Edited by M.I. Holland / Elsevier, Amsterdam, 1992).
Posner. (MIT Press, Cambridge (MA), 1989). [93] A. Stern. The Quantum Brain. (North-Holland /
[80] K.H. Pribram. Brain and Perception. (Lawrence Elsevier, Amsterdam, 1994).
Erlbaum Assoc., Hillsdale (NJ), 1991). [94] N.V. Swindale. The development of topography
[81] I. Prigogine. Why Irreversibility? The Formu- in the visual cortex: a review of models. Network:
lation of Classical and Quantum Mechanics for Comp. in Neural Sys. 7, 161-247 (1996).
Nonintegrable Systems. Int. J. Quantum Chem. [95] J. Šlechta. On a Quantum-Statistical Theory of
53, 105-118 (1995). Pair Interaction Between Memory Traces in the
[82] H. Ritter, T. Martinetz, K. Schulten. Neu- Brain. Informatica 17, 109-115 (1993) (and other
ral Computation and Self-Organizing Maps. ref. within).
(Addison-Wesley, Redwood City, 1992). [96] J. Šlechta. A Study of Structurally Disordered
[83] Neurobiology of Higher Cognitive Functions. Single-Element Magnetic Materials beyond the
Edited by A.B. Scheibel, A.F. Wechsler. (Guil- Homogeneous Molecular Field Approximation.
ford Press, London, 1990). Phys. stat. sol. (b) 67, 595-536 (1975).
[84] W. Schempp. Bohr’s Indeterminacy Principle in [97] J.G. Taylor, F.N. Alavi. Mathematical Analy-
Quantum Holography, Self-adaptive Neural Net- sis of a Competitive Network for Attention. In:
work Architectures, Cortical Self-Organization, Mathematical Approaches to Neural Networks.
Molecular Computers, Magnetic Resonance Edited by J.G. Taylor. (Elsevier, Amsterdam,
Imaging and Solitonic Nanotechnology. Nanobiol. 1993).
2, 109 (1993). [98] V. Tryba, K. Goser. A Modified Algorithm
[85] W. Schempp. Analog VLSI network models, cor- for Self-Organizing Maps Based on Schrödinger
tical linking neural network models, and quantum Equation. In: Artificial Neural Networks, Pro-
holographic neural technology. In: Wavelets and ceed. Int. Workshop IWANN, 1991.
Their Applications. Edited by J.S. Byrnes et al. [99] M. Tušar, J. Zupan, J. Gasteiger. Neural net-
(Kluwer, Amsterdam, 1994, pp. 213-260). works modelling in chemistry. J. Chem. Phys. 89,
[86] W. Schempp. Phase coherent wavelets, Fourier 1517 (1992).
transform magnetic resonance imaging, and syn- [100] G. Vitiello. Coherence and Electromagnetic
chronized time-domain neural networks. Proceed. Fields in Living Matter. Nanobiol. 1, 221-228
Steklov Inst. Math. 3, 323-351 (1995). (1992).
[87] G.I. Shipov. A Theory of Physical Vacuum (chap- [101] Proceedings IJC ”Inform. Sci. ’98” – vol. 2: Pro-
ters 2.4 and 2.5, especially). (book manuscript, ceed. Int. Conf. ”Comput. Intellig. & Neurosci.
Moscow, 1997). ’98”. Edited by P.P. Wang et al. (Assoc. Intellig.
[88] L. Shastri, V. Ajjanagadde. From simple asso- Mach., RTP (NC, USA), 1998, pp. 167-224) (pa-
ciations to systematic reasoning: A connection- pers by Behrman, Vitiello, Ventura, Ohya, Linke-
ist representation of rules, variables and dynamic vich, MacLennan, Gould, Matsuoka, Beck, Kak,
bindings using temporal synchrony. Behav. & Peruš, et al.).
Brain Sci. 16, 417 (1993). [102] A. Zippelius. Statistical Mechanics of Neural Net-
[89] H. Sompolinsky. Statistical Mechanics of Neural works. Physica A 194, 471-481 (1993).
Networks. Phys. Today, pp. 70 (December 1988). [103] J. Zupan. Introduction to Artificial Neural Net-
[90] H. Sompolinsky, M. Tsodyks. Segmentation by work (ANN) Methods: What They Are and How
a Network of Oscillators with Stored Memories. to Use Them. Acta Chimica Slov. 41, 327 (1994).
Neural Computat. 6, 642-657 (1994).