wetware
'Wolfgang Maass
Wolfgang Maass
If you pour water over your PC, the PC will stop working. This is because very late in the history of computing—which started about 500 million years ago (1) —the PC and other devices for information processing were developed that require a dry environment. But these new devices, consisting of hardware and software, have a disadvantage: they do not work as well as the older and more common computational devices that are called nervous systems, or brains, and which consist of wetware. These superior computational devices were made to function in a somewhat salty aqueous solution, apparently because many of the first creatures with a nervous system came from the sea. We still carry an echo of this history of computing in our heads: the neurons in our brain are embedded in an artificial sea-environment, the salty aqueous extracellular fluid which surrounds the neurons in our brain. The close relationship between the wetware in our brain and the wetware in evolutionary much older organisms that still live in the sea is actually quite helpful for research. Neurons in the squid are 100 to 1000 times larger than the neurons in our brain, and therefore easier to study. Nevertheless the equations that Hodgkin and Huxley derived to model the dynamics of the neuron that controls the escape reflex of the squid (for which they received the Nobel prize in 1963), also apply to the neurons in our brain. In this short paper I want to give you a glimpse of this foreign world of computing in wetware.
One of the technical problems that nature had to solve to enabe computation in wetware was how to communicate intermediate results from the computation of one neuron to other neurons, or to output-devices such as muscles. In a PC one sends streams of bits over copper wires. But copper wires were not available a few hundred million years ago, nor do they work as well in a sea-environment. The solution that nature found was the so-called action potential or spike. The spike plays in a brain a similar role to that of a bit in a digital computer: it is the common unit of information in wetware. A spike is a sudden voltage increase for about 1 ms (1 ms = 1/1000 second) that is created at the cell body (soma) of a neuron, more precisely at its trigger zone, and propagated along a lengthy fiber (called axon) that extends from the cell body. This axon corresponds to an insulated copper wire in hardware. The gray matter of your brain contains large amounts of such axons: about 4 km in every cubic millimeter ( = 1 mm3). Axons have numerous branching points (see the axonal tree on the right hand side of Fig. 1), at which most spikes are automatically duplicated, so that they can enter each branch of the axonal tree. In this way a spike from a single neuron can be transmitted to a few thousand other neurons. But in order to move from one neuron to another, the spike has to pass a rather complicated switch, a so-called synapse (marked by a blue triangle in Figure 2, and shown in more detail in Figure 3).
When a spike enters a synapse, it is likely to trigger a complex chain of events that are indicated in Figure 3 (2) : a small vesicle filled with special molecules (“neurotransmitter“) is fused with the cell membrane of the presynaptic terminal, thereby releasing the neurotransmitter into the extracellular fluid. Whenever a neurotransmitter molecule reaches a particular molecular arrangement (a “receptor”) in the cell membrane of the next neuron, it will open a channel in that cell membrane through which charged particles (ions) can enter the next cell. This causes an increase or decrease (depending on the type of channel that is opened and the types of ions that this channel lets through) of the membrane voltage by a few millivolt (1 millivolt = 1/1000 volt). One calls these potential changes EPSPs (excitatory postsynaptic potentials) if they increase the membrane voltage, and otherwise IPSPs (inhibitory postsynaptic potentials). In contrast to the spikes, which all look alike, the size and shape of these postsynaptic potentials depends very much on the particular synapse that causes it. In fact it will also depend on the current “mood" and the recent “experiences" of this synapse, since the postsynaptic potentials are different sizes, depending on the pattern of spikes that have reached the synapse in the past, on the interaction of these spikes with the firing activity of the postsynaptic neuron, and also on other signals that reach the synapse in the form of various molecules (e.g. neurohormones) through the extracellular fluid.
Sometimes people wonder whether it is possible to replace wetware by hardware, to replace for example parts of a brain by silicon chips. This is not so easy because wetware does not consist of fixed computational components, like a silicon chip, that perform the same operation in the same way every day of their working life. Instead the channels and receptors of neurons and synapses move around, disappear, and are replaced by new and possibly different receptors and channels that are continuously reproduced by a living cell in dependence on the individual “experience" of that cell (such as the firing patterns of the pre- and postsynaptic neuron, and the cocktail of biochemical substances that reach the cell through the extracellular fluid). This implies that next year a synapse in your brain is likely to perform its operations quite differently from today, whereas a silicon clone of your brain would be stuck with the “old" synapses from this year.
The postsynaptic potentials created by the roughly 10,000 synapses converging on a single neuron are transmitted by a tree of input wires (“dendritic tree", see Fig. 1) to the trigger zone at the cell body of a neuron. Whenever the sum of these hundreds and thousands of continuously arriving voltage changes reaches the firing threshold there, the neuron will “fire" (a chain reaction orchestrated through the rapid opening of channels in the cell membrane that allow positively charged sodium ions to enter the neuron, thereby increasing the membrane voltage, which causes further channels to open) and send out a spike through its axon. (3) So we are back at our starting point, the spike.
The question is now how a network of neurons can compute with such spikes. Figure 5 presents an illustration of a tiny network consisting of just 3 neurons, which communicate via sequences of spikes (usually referred to as spike trains). It is taken from an animated computer installation which is available online. It allows you to create your own spike train, and watch how the network responds to it. You can also change the strength of the synapses, and thereby simulate (in an extremely simplified manner) processes that take place when the neural system “learns.” (4) But we still do not know how to transmit information via spikes, so let us look at the protocol of a real computation in wetware. In Figure 6 the spike trains emitted by 30 (randomly selected) neurons in the visual area of a monkey brain are shown for a period of 4 seconds. All the information from your senses, all your ideas and thoughts are coded in a similar fashion by spike trains. If you were, for example, to make a protocol of all the visual information which reaches your brain within 4 seconds, you would arrive at a similar figure, but with 1,000,000 rows instead of 30, because the visual information is transmitted from the retina of your eye to your brain by the axons of about 1,000,000 neurons.
Researchers used to think that the only computationally relevant signal in the output of a neuron was the frequency of its firing. But you may notice in Figure 6 that the frequency of firing of a neuron tends to change rapidly, and that the time intervals between the spikes are so irregular that you do not find it easy to estimate the average frequency of firing of a neuron by looking at just 2 or 3 spikes from that neuron. On the other hand our brain can compute quite fast, in about 150 ms, with just 2 or 3 spikes per neuron. This suggests that other features of spike trains must be used by the brain for transmitting information. Recent experimental studies (see for example [Rieke et al., 1997, Koch 1999, Recce 1999]) show that in fact the full spatial and temporal pattern of spikes emitted by neurons is relevant for the message which they are sending to other neurons. Hence it would be more appropriate to compare the output of a collection of neurons with a piece of music played by an orchestra. To recognize such a piece of music it does not suffice to know how often each note is played by each musician. Instead we have to know how the notes of the musicians are embedded in the melody and the pattern of notes played by other musicians. One assumes now that in a similar manner many groups of neurons in the brain code their information through the pattern in which each neuron fires relative to the other neurons in the group. Hence, one may argue that music is a code that is much more closely related to the codes used in your brain than the bit-stream code used by a PC.
The investigation of theoretical and practical possibilities to compute with such spatiotemporal patterns of pulses has led to the creation of a new generation of artificial neural networks, so-called pulsbased neural networks (see [Maass, Bishop] for surveys and recent research results). Such networks are now also appearing in the form of novel electronic hardware [Mead, 1989, Deiss et al, 1999, Murray, 1999]. An interesting feature of these pulsbased neural networks is that they do not require a global synchronisation (like a PC, or a traditional artificial neural network). So they allow time to be used as a new dimension for coding information. In addition they can save a lot of energy, (5) since no clock signal has to be transmitted all the time to all components of the network. One major unsolved problem is the organization of computation in such systems, since the operating system of wetware is still unknown, even for the squid. Hence our current research, jointly with neurobiologists, concentrates on unraveling the organization of computation in neural microcircuits, the lowest level of circuit architecture in the brain (see [Maass, Natschlaeger, and Markram]).
Notes
(1) One could also argue that the history of computing started somewhat earlier, even before any nervous systems existed: 3 to 4 billion years ago when nature discovered information processing via R NA.back
(2) See www.wwnorton.com/gleitman/ch2/tutorials/2tut5.htm> for an online animation.back
(3) See www.wwnorton.com/gleitman/ch2/tutorials/2tut2.htm for an online animation.back
(4) See www.igi.TUGraz.at/demos/index.html. This computer installation was programmed by Thomas Natschlaeger and Harald Burgsteiner, with support from the Steiermaerkische Landesregierung. Detailed explanations and instructions are available online from www.igi.TUGraz.at/maass/118/118.html, see [Maass, 2000b]. Further background information is available online from [Natschlaeger], [Maass, 2000a], [Maass, 2001].back
(5) Wetware consumes much less energy than any hardware that is currently available.. Our brain, which has about as many computational units as a very large supercomputer, consumes just 10 to 20 watts.back
References
DEISS, S. R., DOUGLAS, R. J., AND WHATLEY, A. M.: “A pulse-coded communications infrastructure for neuromorphic systems,” in Maass, W., and Bishop, C., editors, Pulsed Neural Networks. MIT-Press, Cambridge MA,1999
KOCH, C.: “Biophysics of Computation: Information Processing” in Single Neurons. Oxford University Press, Oxford, 1999
KRÜGER, J., AND AIPLE, F.: “Multielectrode investigation of monkey striate cortex: spike train correlations in the infragranular layers“. Neurophysiology. pp 60:798-828. 1988
MAASS, W. (2000a).: “Das menschliche Gehirn – nur ein Rechner?” In Burkard, R. E., Maass, W., and P. Weibel, editors, Zur Kunst des Formalen Denkens, pp 209-233. Passagen Verlag, Vienna, 2000. See #108 on http://www.igi.tugraz.at/maass/publications.html
MAASS, W. (2000b).: “Spike trains – im Rhythmus neuronaler Zellen.” In Kriesche, R., and Konrad, H., editors, Katalog der steirischen Landesausstellung gr2000az. pp 36–42. Springer Verlag. See #118 on www.igi.tugraz.at/maass/publications.html
MAASS, W.: “Paradigms for computing with spiking neurons.” In Leo van Hemmen, editor, Models of Neural Networks, volume 4. Springer Berlin, to appear 2001 See #110 on www.igi.tugraz.at/maass/publications.html
MAASS, W., AND BISHOP, C., editors: Pulsed Neural Networks. MIT-Press, Cambridge, MA, 1999. Paperback, 2001. See www.igi.tugraz.at/maass/PNN.html
MAASS, W., NATSCHLAEGER, T., AND MARKRAM, H.: Real-time computing without stable states: a new framework for neural computation based on perturbations, submitted for publication, 2001
MEAD, C.: Analog VLSI and Neural Systems. Addison-Wesley, Reading, 1989
MURRAY, A. F.: “Pulse-based computation in VLSI neural networks.” In Maass, W., and Bishop, C., editors, Pulsed Neural Networks. MIT-Press, Cambridge MA, 1999
NATSCHLÄGER, T.: “Die dritte Generation von Modellen für neuronale Netzwerke Netzwerke von Spiking Neuronen.” In: Jenseits von Kunst. Passagen Verlag, 1996. See www.igi.tugraz.at/tnatschl/
RECCE, M.: “Encoding information in neuronal activity.” In Maass, W., and Bishop, C., editors, Pulsed Neural Networks. MIT-Press, Cambridge MA, 1999
RIEKE, F., WARLAND, D., BIALEK, W., and DE RUYTER VAN STEVENINCK, R.: SPI KES: Exploring the Neural Code. MIT-Press, Cambridge MA, 1997
|