Part 7 (2/2)

Babbage's conception is quite miraculous when you consider the era in which he lived and worked. However, by the mid-twentieth century, his ideas had been lost in the mists of time (although they were subsequently rediscovered). It was von Neumann who conceptualized and articulated the key principles of the computer as we know it today, and the world recognizes this by continuing to refer to the von Neumann machine as the princ.i.p.al model of computation. Keep in mind, though, that the von Neumann machine continually communicates data between its various units and within these units, so it could not be built without Shannon's theorems and the methods he devised for transmitting and storing reliable digital information.

That brings us to the fourth important idea, which is to go beyond Ada Byron's conclusion that a computer could not think creatively and find the key algorithms employed by the brain and then use these to turn a computer into a brain. Alan Turing introduced this goal in his 1950 paper ”Computing Machinery and Intelligence,” which includes his now-famous Turing test for ascertaining whether or not an AI has achieved a human level of intelligence.

In 1956 von Neumann began preparing a series of lectures intended for the prestigious Silliman lecture series at Yale University. Due to the ravages of cancer, he never delivered these talks nor did he complete the ma.n.u.script from which they were to be given. This unfinished doc.u.ment nonetheless remains a brilliant and prophetic foreshadowing of what I regard as humanity's most daunting and important project. It was published posthumously as The Computer and the Brain The Computer and the Brain in 1958. It is fitting that the final work of one of the most brilliant mathematicians of the last century and one of the pioneers of the computer age was an examination of intelligence itself. This project was the earliest serious inquiry into the human brain from the perspective of a mathematician and computer scientist. Prior to von Neumann, the fields of computer science and neuroscience were two islands with no bridge between them. in 1958. It is fitting that the final work of one of the most brilliant mathematicians of the last century and one of the pioneers of the computer age was an examination of intelligence itself. This project was the earliest serious inquiry into the human brain from the perspective of a mathematician and computer scientist. Prior to von Neumann, the fields of computer science and neuroscience were two islands with no bridge between them.

Von Neumann starts his discussion by articulating the similarities and differences between the computer and the human brain. Given when he wrote this ma.n.u.script, it is remarkably accurate. He noted that the output of neurons was digital-an axon either fired or it didn't. This was far from obvious at the time, in that the output could have been an a.n.a.log signal. The processing in the dendrites leading into a neuron and in the soma neuron cell body, however, was a.n.a.log, and he described its calculations as a weighted sum of inputs with a threshold. This model of how neurons work led to the field of connectionism, which built systems based on this neuron model in both hardware and software. (As I described in the previous chapter previous chapter, the first such connectionist system was created by Frank Rosenblatt as a software program on an IBM 704 computer at Cornell in 1957, immediately after von Neumann's draft lectures became available.) We now have more sophisticated models of how neurons combine inputs, but the essential idea of a.n.a.log processing of dendrite inputs using neurotransmitter concentrations has remained valid.

Von Neumann applied the concept of the universality of computation to conclude that even though the architecture and building blocks appear to be radically different between brain and computer, we can nonetheless conclude that a von Neumann machine can simulate the processing in a brain. The converse does not hold, however, because the brain is not a von Neumann machine and does not have a stored program as such (albeit we can simulate a very simple Turing machine in our heads). Its algorithm or methods are implicit in its structure. Von Neumann correctly concludes that neurons can learn patterns from their inputs, which we have now established are coded in part in dendrite strengths. What was not known in von Neumann's time is that learning also takes place through the creation and destruction of connections between neurons.

Von Neumann presciently notes that the speed of neural processing is extremely slow, on the order of a hundred calculations per second, but that the brain compensates for this through ma.s.sive parallel processing-another un.o.bvious and key insight. Von Neumann argued that each one of the brain's 1010 neurons (a tally that itself was reasonably accurate; estimates today are between 10 neurons (a tally that itself was reasonably accurate; estimates today are between 1010 and 10 and 1011) was processing at the same time. In fact, each of the connections (with an average of about 103 to 10 to 104 connections per neuron) is computing simultaneously. connections per neuron) is computing simultaneously.

Von Neumann's estimates and his descriptions of neural processing are remarkable, given the primitive state of neuroscience at the time. One aspect of his work that I do disagree with, however, is his a.s.sessment of the brain's memory capacity. He a.s.sumes that the brain remembers every input for its entire life. Von Neumann a.s.sumes an average life span of 60 years, or about 2 109 seconds. With about 14 inputs to each neuron per second (which is actually low by at least three orders of magnitude) and with 10 seconds. With about 14 inputs to each neuron per second (which is actually low by at least three orders of magnitude) and with 1010 neurons, he arrives at an estimate of about 10 neurons, he arrives at an estimate of about 1020 bits for the brain's memory capacity. The reality, as I have noted earlier, is that we remember only a very small fraction of our thoughts and experiences, and even these memories are not stored as bit patterns at a low level (such as a video image), but rather as sequences of higher-level patterns. bits for the brain's memory capacity. The reality, as I have noted earlier, is that we remember only a very small fraction of our thoughts and experiences, and even these memories are not stored as bit patterns at a low level (such as a video image), but rather as sequences of higher-level patterns.

As von Neumann describes each mechanism in the brain, he shows how a modern computer could accomplish the same thing, despite their apparent differences. The brain's a.n.a.log mechanisms can be simulated through digital ones because digital computation can emulate a.n.a.log values to any desired degree of precision (and the precision of a.n.a.log information in the brain is quite low). The brain's ma.s.sive parallelism can be simulated as well, given the significant speed advantage of computers in serial computation (an advantage that has vastly expanded over time). In addition, we can also use parallel processing in computers by using parallel von Neumann machines-which is exactly how supercomputers work today.

Von Neumann concludes that the brain's methods cannot involve lengthy sequential algorithms, when one considers how quickly humans are able to make decisions combined with the very slow computational speed of neurons. When a third baseman fields a ball and decides to throw to first rather than to second base, he makes this decision in a fraction of a second, which is only enough time for each neuron to go through a handful of cycles. Von Neumann concludes correctly that the brain's remarkable powers come from all its 100 billion neurons being able to process information simultaneously. As I have noted, the visual cortex makes sophisticated visual judgments in only three or four neural cycles.

There is considerable plasticity in the brain, which enables us to learn. But there is far greater plasticity in a computer, which can completely restructure its methods by changing its software. Thus, in that respect, a computer will be able to emulate the brain, but the converse is not the case.

When von Neumann compared the capacity of the brain's ma.s.sively parallel organization to the (few) computers of his time, it was clear that the brain had far greater memory and speed. By now the first supercomputer to achieve specifications matching some of the more conservative estimates of the speed required to functionally simulate the human brain (about 1016 operations per second) has been built. operations per second) has been built.5 (I estimate that this level of computation will cost $1,000 by the early 2020s.) With regard to memory we are even closer. Even though it was remarkably early in the history of the computer when his ma.n.u.script was written, von Neumann nonetheless had confidence that both the hardware and software of human intelligence would ultimately fall into place, which was his motivation for having prepared these lectures. (I estimate that this level of computation will cost $1,000 by the early 2020s.) With regard to memory we are even closer. Even though it was remarkably early in the history of the computer when his ma.n.u.script was written, von Neumann nonetheless had confidence that both the hardware and software of human intelligence would ultimately fall into place, which was his motivation for having prepared these lectures.

Von Neumann was deeply aware of the increasing pace of progress and its profound implications for humanity's future. A year after his death in 1957, fellow mathematician Stan Ulam quoted him as having said in the early 1950s that ”the ever accelerating progress of technology and changes in the mode of human life give the appearance of approaching some essential singularity in the history of the race beyond which human affairs, as we know them, could not continue.” This is the first known use of the word ”singularity” in the context of human technological history.

Von Neumann's fundamental insight was that there is an essential equivalence between a computer and the brain. Note that the emotional intelligence of a biological human is part of its intelligence. If von Neumann's insight is correct, and if one accepts my own leap of faith that a nonbiological ent.i.ty that convincingly re-creates the intelligence (emotional and otherwise) of a biological human is conscious (see the next chapter next chapter), then one would have to conclude that there is an essential equivalence between a computer-with the right software-and a (conscious) mind. So is von Neumann correct?

Most computers today are entirely digital, whereas the human brain combines digital and a.n.a.log methods. But a.n.a.log methods are easily and routinely re-created by digital ones to any desired level of accuracy. American computer scientist Carver Mead (born in 1934) has shown that we can directly emulate the brain's a.n.a.log methods in silicon, which he has demonstrated with what he calls ”neuromorphic” chips.6 Mead has demonstrated how this approach can be thousands of times more efficient than digitally emulating a.n.a.log methods. As we codify the ma.s.sively repeated neocortical algorithm, it will make sense to use Mead's approach. The IBM Cognitive Computing Group, led by Dharmendra Modha, has introduced chips that emulate neurons and their connections, including the ability to form new connections. Mead has demonstrated how this approach can be thousands of times more efficient than digitally emulating a.n.a.log methods. As we codify the ma.s.sively repeated neocortical algorithm, it will make sense to use Mead's approach. The IBM Cognitive Computing Group, led by Dharmendra Modha, has introduced chips that emulate neurons and their connections, including the ability to form new connections.7 Called ”SyNAPSE,” one of the chips provides a direct simulation of 256 neurons with about a quarter million synaptic connections. The goal of the project is to create a simulated neocortex with 10 billion neurons and 100 trillion connections-close to a human brain-that uses only one kilowatt of power. Called ”SyNAPSE,” one of the chips provides a direct simulation of 256 neurons with about a quarter million synaptic connections. The goal of the project is to create a simulated neocortex with 10 billion neurons and 100 trillion connections-close to a human brain-that uses only one kilowatt of power.

As von Neumann described over a half century ago, the brain is extremely slow but ma.s.sively parallel. Today's digital circuits are at least 10 million times faster than the brain's electrochemical switches. Conversely, all 300 million of the brain's neocortical pattern recognizers process simultaneously, and all quadrillion of its interneuronal connections are potentially computing at the same time. The key issue for providing the requisite hardware to successfully model a human brain, though, is the overall memory and computational throughput required. We do not need to directly copy the brain's architecture, which would be a very inefficient and inflexible approach.

Let's estimate what those hardware requirements are. Many projects have attempted to emulate the type of hierarchical learning and pattern recognition that takes place in the neocortical hierarchy, including my own work with hierarchical hidden Markov models. A conservative estimate from my own experience is that emulating one cycle in a single pattern recognizer in the biological brain's neocortex would require about 3,000 calculations. Most simulations run at a fraction of this estimate. With the brain running at about 102 (100) cycles per second, that comes to 3 10 (100) cycles per second, that comes to 3 105 (300,000) calculations per second per pattern recognizer. Using my estimate of 3 10 (300,000) calculations per second per pattern recognizer. Using my estimate of 3 108 (300 million) pattern recognizers, we get about 10 (300 million) pattern recognizers, we get about 1014 (100 trillion) calculations per second, a figure that is consistent with my estimate in (100 trillion) calculations per second, a figure that is consistent with my estimate in The Singularity Is Near The Singularity Is Near. In that book I projected that to functionally simulate the brain would require between 1014 and 10 and 1016 calculations per second (cps) and used 10 calculations per second (cps) and used 1016 cps to be conservative. AI expert Hans Moravec's estimate, based on extrapolating the computational requirement of the early (initial) visual processing across the entire brain, is 10 cps to be conservative. AI expert Hans Moravec's estimate, based on extrapolating the computational requirement of the early (initial) visual processing across the entire brain, is 1014 cps, which matches my own a.s.sessment here. cps, which matches my own a.s.sessment here.

Routine desktop machines can reach 1010 cps, although this level of performance can be significantly amplified by using cloud resources. The fastest supercomputer, j.a.pan's K Computer, has already reached 10 cps, although this level of performance can be significantly amplified by using cloud resources. The fastest supercomputer, j.a.pan's K Computer, has already reached 1016 cps. cps.8 Given that the algorithm of the neocortex is ma.s.sively repeated, the approach of using neuromorphic chips such as the IBM SyNAPSE chips mentioned above is also promising. Given that the algorithm of the neocortex is ma.s.sively repeated, the approach of using neuromorphic chips such as the IBM SyNAPSE chips mentioned above is also promising.

In terms of memory requirement, we need about 30 bits (about four bytes) for one connection to address one of 300 million other pattern recognizers. If we estimate an average of eight inputs to each pattern recognizer, that comes to 32 bytes per recognizer. If we add a one-byte weight for each input, that brings us to 40 bytes. Add another 32 bytes for downward connections, and we are at 72 bytes. Note that the branching-up-and-down figure will often be much higher than eight, though these very large branching trees are shared by many recognizers. For example, there may be hundreds of recognizers involved in recognizing the letter ”p.” These will feed up into thousands of such recognizers at this next higher level that deal with words and phrases that include ”p.” However, each ”p” recognizer does not repeat the tree of connections that feeds up to all of the words and phrases that include ”p”-they all share one such tree of connections. The same is true of downward connections: A recognizer that is responsible for the word ”APPLE” will tell all of the thousands of ”E” recognizers at a level below it that an ”E” is expected if it has already seen ”A,” ”P,” ”P,” and ”L.” That tree of connections is not repeated for each word or phrase recognizer that wants to inform the next lower level that an ”E” is expected. Again, they are shared. For this reason, an overall estimate of eight up and eight down on average per pattern recognizer is reasonable. Even if we increase this particular estimate, it does not significantly change the order of magnitude of the resulting estimate.

With 3 108 (300 million) pattern recognizers at 72 bytes each, we get an overall memory requirement of about 2 10 (300 million) pattern recognizers at 72 bytes each, we get an overall memory requirement of about 2 1010 (20 billion) bytes. That is actually a quite modest number that routine computers today can exceed. (20 billion) bytes. That is actually a quite modest number that routine computers today can exceed.

These estimates are intended only to provide rough estimates of the order of magnitude required. Given that digital circuits are inherently about 10 million times faster than the biological neocortical circuits, we do not need to match the human brain for parallelism-modest parallel processing (compared with the trillions-fold parallelism of the human brain) will be sufficient. We can see that the necessary computational requirements are coming within reach. The brain's rewiring of itself-dendrites are continually creating new synapses-can also be emulated in software using links, a far more flexible system than the brain's method of plasticity, which as we have seen is impressive but limited.

The redundancy used by the brain to achieve robust invariant results can certainly be replicated in software emulations. The mathematics of optimizing these types of self-organizing hierarchical learning systems is well understood. The organization of the brain is far from optimal. Of course it didn't need to be-it only needed to be good enough to achieve the threshold of being able to create tools that would compensate for its own limitations.

Another restriction of the human neocortex is that there is no process that eliminates or even reviews contradictory ideas, which accounts for why human thinking is often ma.s.sively inconsistent. We have a weak mechanism to address this called critical thinking, but this skill is not practiced nearly as often as it should be. In a software-based neocortex, we can build in a process that reveals inconsistencies for further review.

It is important to note that the design of an entire brain region is simpler than the design of a single neuron. As discussed earlier, models often get simpler at a higher level-consider an a.n.a.logy with a computer. We do need to understand the detailed physics of semiconductors to model a transistor, and the equations underlying a single real transistor are complex. A digital circuit that multiples two numbers requires hundreds of them. Yet we can model this multiplication circuit very simply with one or two formulas. An entire computer with billions of transistors can be modeled through its instruction set and register description, which can be described on a handful of written pages of text and formulas. The software programs for an operating system, language compilers, and a.s.semblers are reasonably complex, but modeling a particular program-for example, a speech recognition program based on hierarchical hidden Markov modeling-may likewise be described in only a few pages of equations. Nowhere in such a description would be found the details of semiconductor physics or even of computer architecture.

A similar observation holds true for the brain. A particular neocortical pattern recognizer that detects a particular invariant visual feature (such as a face) or that performs a bandpa.s.s filtering (restricting input to a specific frequency range) on sound or that evaluates the temporal proximity of two events can be described with far fewer specific details than the actual physics and chemical relations controlling the neurotransmitters, ion channels, and other synaptic and dendritic variables involved in the neural processes. Although all of this complexity needs to be carefully considered before advancing to the next higher conceptual level, much of it can be simplified as the operating principles of the brain are revealed.

CHAPTER 9

THOUGHT EXPERIMENTS ON THE MIND

Minds are simply what brains do.-Marvin Minsky, The Society of Mind The Society of Mind

When intelligent machines are constructed, we should not be surprised to find them as confused and as stubborn as men in their convictions about mind-matter, consciousness, free will, and the like.-Marvin Minsky, The Society of Mind The Society of Mind

Who Is Conscious?The real history of consciousness starts with one's first lie.-Joseph Brodsky

Suffering is the sole origin of consciousness.-Fyodor Dostoevsky, Notes from Underground Notes from Underground

There is a kind of plant that eats organic food with its flowers: when a fly settles upon the blossom, the petals close upon it and hold it fast till the plant has absorbed the insect into its system; but they will close on nothing but what is good to eat; of a drop of rain or a piece of stick they will take no notice. Curious! that so unconscious a thing should have such a keen eye to its own interest. If this is unconsciousness, where is the use of consciousness?-Samuel Butler, 1871

We have been examining the brain as an ent.i.ty that is capable of certain levels of accomplishment. But that perspective essentially leaves our selves selves out of the picture. We appear to live in our brains. We have subjective lives. How does the objective view of the brain that we have discussed up until now relate to our own feelings, to our sense of being the person having the experiences? out of the picture. We appear to live in our brains. We have subjective lives. How does the objective view of the brain that we have discussed up until now relate to our own feelings, to our sense of being the person having the experiences?

British philosopher Colin McGinn (born in 1950) writes that discussing ”consciousness can reduce even the most fastidious thinker to blabbering incoherence.” The reason for this is that people often have unexamined and inconsistent views on exactly what the term means.

Many observers consider consciousness to be a form of performance-for example, the capacity for self-reflection, that is, the ability to understand one's own thoughts and to explain them. I would describe that as the ability to think about one's own thinking. Presumably, we could come up with a way of evaluating this ability and then use this test to separate conscious things from unconscious things.

However, we quickly get into trouble in trying to implement this approach. Is a baby conscious? A dog? They're not very good at describing their own thinking process. There are people who believe that babies and dogs are not conscious beings precisely because they cannot explain themselves. How about the computer known as Watson? It can be put into a mode where it actually does explain how it came up with a given answer. Because it contains a model of its own thinking, is Watson therefore conscious whereas the baby and the dog are not?

Before we proceed to pa.r.s.e this question further, it is important to reflect on the most significant distinction relating to it: What is it that we can ascertain from science, versus what remains truly a matter of philosophy? One view is that philosophy is a kind of halfway house for questions that have not yet yielded to the scientific method. According to this perspective, once science advances sufficiently to resolve a particular set of questions, philosophers can then move on to other concerns, until such time that science resolves them also. This view is endemic where the issue of consciousness is concerned, and specifically the question ”What and who is conscious?”

Consider these statements by philosopher John Searle: ”We know that brains cause consciousness with specific biological mechanisms.... The essential thing is to recognize that consciousness is a biological process like digestion, lactation, photosynthesis, or mitosis.... The brain is a machine, a biological machine to be sure, but a machine all the same. So the first step is to figure out how the brain does it and then build an artificial machine that has an equally effective mechanism for causing consciousness.”1 People are often surprised to see these quotations because they a.s.sume that Searle is devoted to protecting the mystery of consciousness against reductionists like Ray Kurzweil. People are often surprised to see these quotations because they a.s.sume that Searle is devoted to protecting the mystery of consciousness against reductionists like Ray Kurzweil.

The Australian philosopher David Chalmers (born in 1966) has coined the term ”the hard problem of consciousness” to describe the difficulty of pinning down this essentially indescribable concept. Sometimes a brief phrase encapsulates an entire school of thought so well that it becomes emblematic (for example, Hannah Arendt's ”the ba.n.a.lity of evil”). Chalmers's famous formulation accomplishes this very well.

When discussing consciousness, it becomes very easy to slip into considering the observable and measurable attributes that we a.s.sociate with being conscious, but this approach misses the very essence of the idea. I just mentioned the concept of metacognition-the idea of thinking about one's own thinking-as one such correlate of consciousness. Other observers conflate emotional intelligence or moral intelligence with consciousness. But, again, our ability to express a loving sentiment, to get the joke, or to be s.e.xy are simply types of performances-impressive and intelligent perhaps, but skills that can nonetheless be observed and measured (even if we argue about how to a.s.sess them). Figuring out how the brain accomplishes these sorts of tasks and what is going on in the brain when we do them const.i.tutes Chalmers's ”easy” question of consciousness. Of course, the ”easy” problem is anything but and represents perhaps the most difficult and important scientific quest of our era. Chalmers's ”hard” question, meanwhile, is so hard that it is essentially ineffable.

<script>