- Welcome to MIT course 6S099, Artificial General Intelligence. Today we have Ray Kurzweil. He is one of the world's leading inventors, thinkers, and futurists, with a 30-year track record of accurate predictions. Called the restless genius by the Wall Street Journal and the ultimate thinking machine by Forbes magazine. He was selected as one of the top entrepreneurs by Inc.
Magazine, which described him as the rightful heir to Thomas Edison. PBS selected him as one of the 16 revolutionaries who made America. Ray was the principal investigator of the first CCD flatbed scanner, the first omni-font optical character recognition, the first point-to-speech reading machine for the blind, the first text-to-speech synthesizer, the first music synthesizer capable of recreating the grand piano and other orchestral instruments, and the first commercially marketed large vocabulary speech recognition.
Among Ray's many honors, he received a Grammy Award for Outstanding Achievements in Music Technology. He is the recipient of the National Medal of Technology, was inducted into the National Inventors Hall of Fame, holds 21 honorary doctorates and honors from three US presidents. Ray has written five national best-selling books, including the New York Times bestsellers The Singularity is Near from 2005, and How to Create a Mind from 2012.
He is co-founder and chancellor of Singularity University and a director of engineering at Google, heading up a team developing machine intelligence and natural language understanding. Please give Ray a warm welcome. (audience applauding) - It's good to be back. I've been in this lecture hall many times and walked the infinite corridor.
I came here as an undergraduate in 1965. Within a year of my being here, they started a new major called computer science. It did not get its own course number. It's 6.1. Even biotechnology recently got its own course number. How many of you are CS majors? Okay, how many of you do work in deep learning?
How many of you have heard of deep learning? I came here first in 1962 when I was 14. I became excited about artificial intelligence. It had only gotten its name six years earlier, the 1956 Dartmouth Conference by Marvin Minsky and John McCarthy. So I wrote Minsky a letter. There was no email back then.
And he invited me up. He spent all day with me as if he had nothing else to do. He was a consummate educator. And the AI field had already bifurcated into two warring camps, the symbolic school which Minsky was associated with and the connectionist school was not widely known.
In fact, I think it's still not widely known that Minsky actually invented the neural net in 1953. But he had become negative about it, largely 'cause there was a lot of hype that these giant brains could solve any problem. So the first popular neural net, the perceptron, was being promulgated by Frank Rosenblatt at Cornell.
So Minsky said, "Oh, where are you going now?" And I said to see Rosenblatt at Cornell. He said, "Don't bother doing that." And I went there and Rosenblatt was touting the perceptron that it ultimately would be able to solve any problem. So I brought some printed letters that had the camera and it did a perfect job of recognizing them as long as they were carrier 10, different type style didn't work at all.
And he said, "But don't worry. "We can take the output of the perceptron "and feed it as the input to another perceptron "and take the output of that and feed it to a third layer. "And as we add more layers, "it'll get smarter and smarter and generalized." And I said, "That's interesting.
"Have you tried that?" Well, no, but it's high on our research agenda. Things did not move quite as quickly back then as they do now. He died nine years later, never having tried that idea. Turns out to be remarkably prescient. I mean, he never tried multi-layer neural nets and all the excitement that we see now about deep learning comes from a combination of two things, many layer neural nets and the law of accelerating returns, which I'll get to a little bit later, which is basically the exponential growth of computing so that we can run these massive nets and handle massive amounts of data.
It would be decades before that idea was tried. Several decades later, three-level neural nets were tried. They were a little bit better. They could deal with multiple type styles, still weren't very flexible. Now, it's not hard to add other layers. It's a very straightforward concept. There was a math problem, the disappearing gradient or the exploding gradient, which I'm sure many of you are familiar with.
Basically, you need to take maximum advantage of the range of values in the gradients, not let them explode or disappear and lose the resolution. That's a fairly straightforward mathematical transformation. With that insight, we could now go to 100-layer neural nets. And that's behind sort of all the fantastic gains that we've seen recently.
AlphaGo trained on every online game and then became a fair Go player. It then trained itself by playing itself and soared past the best human. AlphaGo Zero started with no human input at all. Within hours of iteration, soared past AlphaGo, also soared past the best chess programs. They had another innovation.
Basically, you need to evaluate the quality of the board at each point, and they used another 100-layer neural net to do that evaluation. So, there's still a problem in the field. Which is, there's a motto that life begins at a billion examples. One of the reasons I'm at Google is we have a billion examples.
For example, there's pictures of dogs and cats that are labeled, so you got a picture of a cat and it says cat, and then you can learn from it, and you need a lot of them. AlphaGo trained on a million online moves. That's how many we had of Master Games.
And that only created a sort of fair Go player, a good amateur could defeat it. So, they worked around that in the case of Go by basically generating an infinite amount of data by having the system play itself. Had a chat with Demis Hassabis. What kind of situations can you do that with?
You have to have some way of simulating the world. So, Go or Chess are, even though Go is considered a difficult game, the definition of it exists on one page. So, you can simulate it. That applies to math. I mean, math axioms can be contained on a page or two.
It's not very complicated. Gets more difficult when you have real life situations, like biology. So, we have biological simulators, but the simulators aren't perfect. So, learning from the simulators will only be as good as the simulators. That's actually the key to being able to do deep learning on biology.
Autonomous vehicles, you need real life data. So, the Waymo systems have gone three and a half million miles. That's enough data to then create a very good simulator. So, the simulator is really quite realistic because they had a lot of real world experience and they've gone a billion miles in the simulator.
But we don't always have that opportunity to either create the data or have the data around. Humans can learn from a small number of examples. Your significant other, your professor, your boss, your investor can tell you something once or twice and you might actually learn from that. Some humans have been reported to do that.
And that's kind of the remaining advantage of humans. Now, there's actually no back propagation in the human brain. It doesn't use deep learning. It uses a different architecture. That same year, in 1962, I wrote a paper, How I Thought the Human Brain Worked. There was actually very little neuroscience to go on.
There was one neuroscientist, Vernon Mountcastle, that had something relevant to say, which is he did... I mean, there was the common wisdom at the time, and there's still a lot of neuroscientists that say this, that we have all these different regions of the brain, they do different things, they must be different.
There's V1 in the back of the head where the optic nerve spills into, that can tell that that's a curved line, that's a straight line, does these simple feature extractions on visual images. That's actually a large part of the neocortex. There's a fusiform gyrus up here, which can recognize faces.
We know that because if it gets knocked out through injury or stroke, people can't recognize faces. They will learn it again with a different region of the neocortex. There's the famous frontal cortex, which does language and poetry and music. So these must work on different principles. He did autopsies on the neocortex and all these different regions and found they all looked the same.
They had the same repeating pattern, same interconnections. He said neocortex is neocortex. So I had that hint. Otherwise, I could actually observe human brains in action, which I did from time to time. And there's a lot of hints that you can get that way. For example, if I ask you to recite the alphabet, you actually don't do it from A to Z, you do it as a sequence of sequences, A, B, C, D, E, F, G, H, I, J, K.
So we learn things as forward sequences of sequences. Forward, because if I ask you to recite the alphabet backwards, you can't do it unless you learn that as a new sequence. So these are all interesting hints. I wrote a paper that the neocortex is organized as a hierarchy of modules, and each module can learn a simple pattern.
And that's how I got to meet President Johnson. And that initiated a half century of thinking about this issue. I came to MIT to study with Marvin Minsky. Actually, I came for two reasons. One, that Minsky became my mentor, which was a mentorship that lasted for over 50 years.
The fact that MIT was so advanced, it actually had a computer, which the other colleges I considered didn't have. It was an IBM 7094, 32K of 36-bit words, so it's 150K of core storage, two microsecond cycle time, two cycles for instruction, so a quarter of a MIP. And thousands of students and professors shared that one machine.
In 2012, I wrote a book about this thesis. It's now actually an explosion of neuroscience evidence to support it. The European Brain Reverse Engineering Project has identified a repeating module of about 100 neurons. It's repeated 300 million times, so it's about 30 billion neurons in the neocortex. The neocortex is the outer layer of the brain.
That's the part where we do our thinking. And they can see in each module, axons coming in from another module. And then the output, the single output axon of that module goes as the input to another module. So we can see it organized as a hierarchy. It's not a physical hierarchy.
The hierarchy comes from these connections. The neocortex is a very thin structure. It's actually one module thick. There's six layers of neurons, but it constitutes one module. And we can see that it learns a simple pattern. And for various reasons, I cite in the book, the pattern recognition model it's using is basically a hidden Markov model.
How many of you have worked with Markov models? Okay. That's usually no hands go up when I ask that question. But a Markov model is not, it is learned, but it's not back propagation. It can learn local features. So it's very good for speech recognition. And the speech recognition work I did in the 80s used these Markov models that became the standard approach because it can deal with local variations.
So the fact that a vowel is stretched, you can learn that in a Markov model. It doesn't learn long distance relationships. That's handled by the hierarchy. And something we don't fully understand yet is exactly how the neocortex creates that hierarchy. But we have figured out how it can connect this module to this module.
Does it then grow? I mean, there's no virtual communication or wireless communication. It's an actual connection. So does it grow an axon from one place to another, which could be inches apart? Actually, all these connections are there from birth, like the streets and avenues of Manhattan. There's vertical and horizontal connections.
So if it decides and how it makes that decision is still not fully understood, but it wants to connect this module to this module, there's already a vertical, horizontal and a vertical connection. It just activates them. We can actually see that now. And it can see that happening in real time on non-invasive brain scans.
So there's a tremendous amount of evidence that's in fact, the neocortex is a hierarchy of modules that can learn, each module learns a simple sequential pattern. And even though the patterns we perceive don't seem like sequences, they may seem three-dimensional or even more complicated, they are in fact represented as sequences, but the complexity comes in with the hierarchy.
So the neocortex emerged 200 million years ago with mammals. All mammals have a neocortex. That's one of the distinguishing features of mammals. These first mammals were small. They were rodents, but they were capable of a new type of thinking. Other non-mammalian animals had fixed behaviors, but those fixed behaviors were very well adapted for their ecological niche.
But these new mammals could invent a new behavior. So creativity and innovation was one feature of the neocortex. So a mouse is escaping a predator, its usual escape path is blocked, it will invent a new behavior to deal with it. Probably wouldn't work, but if it did work, it would remember it and it would have a new behavior.
And that behavior could spread virally through the community. Another mouse watching this would say to itself, hmm, that was really clever going around that rock, I'm gonna remember to do that. And it would have a new behavior. Didn't help these early mammals that much because as I say, the non-mammalian animals were very well adapted to their niches and nothing much happened for 135 million years.
But then 65 million years ago, something did happen. There was a sudden violent change to the environment. We now call it the Cretaceous extinction event. There's been debate as to whether it was a meteor or an asteroid, I mean a meteor or a volcanic eruption. The asteroid or meteor hypothesis is in the ascendancy.
But if you dig down to an area of rock reflecting 65 million years ago, the geologists will explain that it shows a very violent sudden change to the environment. And we see it all around the globe. So it was a worldwide phenomenon. The reason we call it an extinction event is that's when the dinosaurs went extinct.
That's when 75% of all the animal and plant species went extinct. And that's when mammals overtook their ecological niche. So to anthropomorphize biological evolution said to itself, this neocortex is pretty good stuff and it began to grow it. So now mammals got bigger, their brains got bigger at an even faster pace, taking up a larger fraction of their body.
The neocortex got bigger even faster than that and developed these curvatures that are distinctive of a primate brain basically to increase its surface area. But if you stretched it out, the human neocortex is still a flat structure. It's about the size of a table napkin, just as thin. And it basically created primates which became dominant in their ecological niche.
Then something else happened 2 million years ago. Biological evolution decided to increase the neocortex further and increase the size of the enclosure and basically filled up the frontal cortex with our big skulls with more neocortex. And up until recently it was felt, as I said, that this was, the frontal cortex was different 'cause it does these qualitatively different things.
But we now realize that it's really just additional neocortex. So remember what we did with it. We're already doing a very good job of being primates. So we put it at the top of the neocortical hierarchy and we increased the size of the hierarchy. It was maybe 20% more neocortex, but it doubled or tripled the number of levels 'cause as you go up the hierarchy, it's kind of like a pyramid.
There's fewer and fewer modules. And that was the enabling factor for us to invent language and art, music. Every human culture we've ever discovered has music. No primate culture really has music. There's debate about that, but it's really true. Invention, technology. Technology required another evolutionary adaptation, which is this humble appendage here.
No other animal has that. If you look at a chimpanzee, it looks like they have a similar hand, but the thumb is actually down here. Doesn't work very well if you watch them trying to grab a stick. So we could imagine creative solutions. Yeah, I could take that branch and strip off the leaves and put a point on it.
We could actually carry out these ideas and create tools and then use tools to create new tools and started a whole other evolutionary process of tool making. And that all came with the neocortex. So Larry Page read my book in 2012 and liked it. So I met with him and asked him for an investment in a company I'd started actually a couple of weeks earlier to develop those ideas commercially 'cause that's how I went about things as a serial entrepreneur.
And he said, "Well, we'll invest, "but let me give you a better idea. "Why don't you do it here at Google? "We have a billion pictures of dogs and cats "and we've got a lot of other data "and lots of computers and lots of talent, "all of which is true." And I says, "Well, I don't know.
"I just started this company to develop this." He says, "Well, buy your company." And I said, "How are you gonna value a company "that hasn't done anything? "It just started a couple of weeks ago." And he said, "We can value anything." So I took my first job five years ago and I've been basically applying this model, this hierarchical model to understanding language, which I think really is the holy grail of AI.
I think Turing was correct in designating basically text communication as what we now call a Turing-complete problem that requires, there's no simple NLP tricks that you can apply to pass a valid Turing test with an emphasis on the word valid. Mitch Caper and I had a six-month debate on what the rules should be 'cause if you read Turing's 1950 paper, he describes this in a few paragraphs and doesn't really describe how to go about it.
But if it's a valid Turing test, meaning it's really convincing you through interrogation and dialogue that it's a human, that requires a full range of human intelligence. And I think that test has stood the test of time. We're making very good progress on that. I mean, just last week, you may have read that two systems passed a paragraph comprehension test.
It's really very impressive. When I came to Google, we were trying to pass these paragraph comprehension tests. We aced the first grade test. Second grade test, we kind of got average performance. And the third grade test had too much inference. Already you had to know some common sense knowledge as it's called and make implications of things that were in different parts of the paragraph.
And there's too much inference and it really didn't work. So this is now adult level, just slightly surpassed average human performance. But we've seen that once something, an AI does something at average human levels, it doesn't take long for it to surpassed average human levels. I think it'll take longer in language than it did in sort of simple games like Go, but it's actually very impressive that it surpasses now average human performance.
It's used in LSTM, long, short temporal memory. But if you look at the adult test, in order to answer these questions, it has to put together inferences and implications of several different things in the paragraph with some common sense knowledge that's not explicitly stated. So that's, I think, a pretty impressive milestone.
So I've been developing, I've got a team of about 45 people, and we've been developing this hierarchical model. We don't use Markov models 'cause we can use deep learning for each module. And so we create an embedding for each word and we create an embedding for each sentence. This, we have a, I can talk about it 'cause we have a published paper on it.
It can take into consideration context. If you use Smart Reply on, if you use Gmail on your phone, you'll see it gives you three suggestions for responses. That's called Smart Reply. They're simple suggestions, but it has to actually understand perhaps a complicated email. And the quality of the suggestions is really quite good, quite on point.
That's for my team using this kind of hierarchical model. So instead of Markov models, it uses embeddings 'cause we can use backpropagation, we might as well use it. But I think what's missing from deep learning is this hierarchical aspect of understanding 'cause the world is hierarchical. That's why evolution developed a hierarchical brain structure to understand the natural hierarchy in the world.
And there's several problems with big, deep neural nets. One is the fact that you really do need a billion examples and we don't, sometimes we can generate them as in the case of Go, or if we have a really good simulator as in the case of autonomous vehicles, not quite the case yet in biology.
Very often you don't have a billion examples. We certainly have billions of examples of language, but they're not annotated. And how would you annotate it anyway with more language that we can't understand in the first place? So it's kind of a chicken and an egg problem. So I believe this hierarchical structure is needed.
Another criticism of deep neural nets is they don't explain themselves very well. It's a big black box that gives you pretty remarkable answers. I mean, in the case of these games, Demis described it's playing in both Go and chess as almost an alien intelligence 'cause we do things that were shocking to human experts like sacrificing a queen and a bishop at the same time or in close succession, which shocked everybody, but then went on to win, or early in a Go game, putting a piece at the corner of the board, which is kind of crazy to most experts 'cause you really wanna start controlling territory.
And yet on reflection, that was the brilliant move that enabled it to win that game. But it doesn't really explain how it does these things. So if you have a hierarchy, it's much better at explaining it 'cause you could look at the content of the modules in the hierarchy and they'll explain what they're doing.
And just to end on the first application of applying this to health and medicine, this will get into high gear and we're gonna really see us break out of the linear extension to longevity that we've experienced. I believe we're only about a decade away from longevity escape velocity. We're adding more time than is going by, not just to infant life expectancy, but to your remaining life expectancy.
I think if someone is diligent, they can be there already. I think I've at longevity escape velocity. Now, a word on what life expectancy means. It used to be assumed that not much would happen. So whatever your life expectancy is with or without scientific progress, it really didn't matter.
Now it matters a lot. So life expectancy really means, how long would you live? What's the, in terms of a statistical likelihood, if there were not continued scientific progress? But that's a very inaccurate assumption. The scientific progress is extremely rapid. I mean, just as an AI in biotech, there are advances now every week.
It's quite stunning. Now, you could have a computed life expectancy, let's say 30 years, 50 years, 70 years from now, you could still be hit by the proverbial bus tomorrow. We're working on that with self-driving vehicles. But we'll get to a point, I think if you're diligent, you can be there now in terms of basically advancing your own statistical life expectancy, at least to keep pace with the passage of time.
I think it will be there for most of the population, at least if they're diligent within about a decade. So if you can hang in there, we may get to see the remarkable century ahead. Thank you very much. (audience applauding) - A question, please raise your hand, we'll get you a mic.
- Hi, so you mentioned both neural network models and symbolic models. And I was wondering how far have you been thinking about combining these two approaches, creating a symbiosis between neural models and symbolic ones? - I don't think we wanna use symbolic models as they've been used. How many are familiar with the Psych Project?
That was a very diligent effort in Texas to define all of common sense reasoning. And it kind of collapsed on itself. And became impossible to debug 'cause you'd fix one thing and it would break three other things. That complexity ceiling has become typical of trying to define things through logical rules.
Now it does seem that humans can understand logical rules. We have logical rules written down for things like law and game playing and so on. But you can actually define a connectionist system to have such a high reliability on a certain type of action that it looks like it's a symbolic rule even though it's represented in a connectionist way.
And connection systems can both capture the soft edges 'cause many things in life are not sharply defined. They can also generate exceptions. So you don't wanna sacrifice your queen in chess except certain situations that might be a good idea. So you can capture that kind of complexity. So we do wanna be able to learn from accumulated human wisdom that looks like it's symbolic.
But I think we'll do it with a connectionist system. But again, I think that connectionist systems should develop a sense of hierarchy and not just be one big massive neural net. - So I understand how we wanna use the neocortex to extract useful stuff and commercialize that. But I'm wondering how our middle brain and the organs that are below the neocortex will be useful for turning that into what you wanna do.
- Well, the cerebellum is an interesting case in point. It actually has more neurons than the neocortex. And it's used to govern most of our behavior. Some things, if you write a signature, that's actually controlled by the cerebellum. So a simple sequence is stored in the cerebellum. But there's not any reasoning to it.
It's basically a script. And most of our movement now has actually been migrated from the cerebellum to the neocortex. Cerebellum is still there. Some people, entire cerebellum is destroyed through disease. They still function fairly normally. Their movement might be a little erratic 'cause our movement is largely controlled by the neocortex.
But some of the subtlety is a kind of pre-programmed script. And so they'll look a little clumsy, but they actually function okay. A lot of other areas of the brain control autonomic functions like breathing. But our thinking really is controlled by the neocortex. In terms of mastering intelligence, I think the neocortex is the brain region we wanna study.
- I'm curious what you think might happen after the singularity is reached in terms of this exponential growth of information. Yeah, do you think it will continue or will there be a whole paradigm shift? What do you predict? - Well, in the singularity's near, I talk about the atomic limits.
Based on molecular computing as we understand it, and it can actually go well past 2045 and actually go to trillions of trillions of times greater computational capacity than we have today. So I don't see that stopping any time soon and will go way beyond what we can imagine. And it becomes an interesting discussion what the impact on human civilization will be.
So to take a maybe slightly more mundane issue that comes up is, oh, it's gonna eliminate most jobs or all jobs. The point I make is it's not the first time in human history we've done that. How many jobs circa 1900 exist today? And that was the feeling of the Luddites, which was an actual society that formed in 1800 after the automation of the textile industry in England.
They looked at all these jobs going away and felt, oh, employment's gonna be just limited to an elite. Indeed, those jobs did go away, but new jobs were created. So if I were a prescient futurist in 1900, I would say, well, 38% of you work on farms and 25% work in factories.
That's 2/3 of the working force. But I predict by 2015, 115 years from now, it's gonna be 2% on farms and 9% in factories and everybody would go, oh my God, we're gonna be out of work. And I said, well, don't worry. For all these jobs we eliminate through automation, we're gonna invent new jobs.
And people say, oh, really, what new jobs? And I'd say, well, I don't know. We haven't invented them yet. That's the political problem. We can see jobs very clearly going away fairly soon, like driving a car or a truck. And the new jobs haven't been invented. I mean, just look at the last five or six years.
A lot of the increase in employment has been through mobile app-related types of ways of making money that just weren't contemplated even six years ago. If I really prescient, I would say, well, you're gonna get jobs creating mobile apps and websites and doing data analytics and self-driving cars. Cars, what's a car?
Nobody would have any idea what I'm talking about. Now, the new job, some people say, yeah, we created new jobs, but it's not as many. Actually, we've gone from 24 million jobs in 1900 to 142 million jobs today, from 30% of the population to 45% of the population. The new jobs pay 11 times as much in constant dollars.
And they're more interesting. I mean, as I talk to people starting out their career now, they really want a career that gives them some life definition and purpose and gratification. We're moving up Maslow's hierarchy. 100 years ago, you were happy if you had a backbreaking job that put food on your family's table.
And we couldn't do these new jobs without enhancing our intelligence. So we've been doing that, well, for most of the last 100 years through education. We've expanded K through 12 in constant dollars tenfold. We've gone from 38,000 college students in 1870 to 15 million today. More recently, we have brain extenders.
They're not yet connected directly in our brain, but they're very close at hand. When I was here at MIT, I had to take my bicycle across campus to get to the computer and show an ID to get in the building. Now we carry them in our pockets and on our belts.
They're gonna go inside our bodies and brains. I think that's not a really important distinction. So we're basically gonna be continuing to enhance our capability through merging with AI. And that's the, I think, ultimate answer to the kind of dystopian view we see in future movies where it's the AI versus a brave band of humans for control of humanity.
We don't have one or two AIs in the world today. We have several billion, three billion smartphones at last count. It'll be six billion in just a couple of years according to the projections. So we're already deeply integrated with this. And I think that's gonna continue. And it's gonna continue to do things which you can't even imagine today.
Just as we are doing today things we couldn't imagine even 20 years ago. - You showed many graphs that go through exponential growth but I haven't seen one that isn't. So I would be very interested in hearing-- - You haven't seen one that, what? - That is not exponential.
So tell me about regions that you've investigated that have not seen exponential growth and why do you think that's the case? - Well, price performance and capacity of information technology invariably follows exponential. When it impacts human society it can be linear. So for example, the growth of democracy has been linear but still pretty steady.
You could count the number of democracies on the fingers of one hand a century ago. Two centuries ago you could count the number of democracies in the world on the fingers of one finger. Now there are dozens of them and it's become kind of a consensus that that's how we should be governed.
So the, and I attribute all this to the growth in information technology, communication in particular for progression of social and cultural institutions. But information technology, because it ultimately depends on a vanishingly small energy and material requirement, grows exponentially and will for a long time. There was recently a criticism that, well, chess scores have, it's actually a remarkably straight linear progression.
So humans, I think it's like 2,800 and it just soared past that in 1997 with Deep Blue and it's kept going. And remarkably straight and saying, well, this is linear, not exponential. But the chess score is a logarithmic measurement. So it really is exponential progression. - So philosophers like to think a lot about the meaning of things, especially in the 20th century.
So for instance, Martin Heidegger gave a couple of speeches and lectures on the relationship of human society to technology and he particularly distinguished between the mode of thinking which is calculating thinking and a mode of thinking which is reflective thinking or meditative thinking. And he posed this question, what is the meaning and purpose of technological development?
And he couldn't find an answer. He recommended to remain open to what he called, he called this an openness to the mystery. I wonder whether you have any thoughts on this. Is there a meaning of purpose to technological development and is there a way for us humans to access that meaning?
- Well, we started using technology to shore up weaknesses in our own capabilities. So physically, I mean, who here could build this building? So we've leveraged the power of our muscles with machines. And we're in fact very bad at doing things that the simplest computers can do, like factor numbers or even just multiply to eight digit numbers.
Computers can do that trivially, we can't do it. So we originally started using computers to make up for that weakness. I think the essence of what I've been writing about is to master the unique strengths of humanity, creating loving expressions in poetry and music and the kinds of things we associate with the better qualities of humanity with machines.
That's the true promise of AI. We're not there yet, but we're making pretty stunning progress. Just in the last year, there's so many milestones that are really significant, including in language. But I think of technology as an expression of humanity. It's part of who we are. And the human species is already a biological technological civilization.
And it's part of who we are. And AI is part of humans. So AI is human and it's part of the technological expression of humanity. And we use technology to extend our reach. I couldn't reach that fruit at that higher branch a thousand years ago. So we invented a tool to extend our physical reach.
And we now extend our mental reach. We can access all of human knowledge with a few keystrokes. And we're gonna make ourselves literally smarter by merging with AI. - Hi, first of all, honored to hear you speak here. So I first read The Singularity as near nine years ago or so.
And it changed the way I thought entirely. But something I think it caused me to over steeply discount was tail risk in geopolitics, in systems that span the entire globe. And my concern is that there is obviously the possibility of tail risk, existential level events, swamping all of these trends that are otherwise war proof, climate proof, you name it.
So my question for you is what steps do you think we can take in designing engineered systems, in designing social and economic institutions to kind of minimize our exposure to these tail risks and survive to make it to a beautiful mind filled future? - Yeah, well, the world was first introduced to a human made existential risk when I was in elementary school, we would have these civil defense drills, get under our desk and put our hands behind our head to protect us from a thermonuclear war.
And it worked, we made it through. But that was really the first introduction to an existential risk. And those weapons are still there, by the way, and they're still on a hair trigger. And they don't get that much attention. There's been a lot of discussion, much of which I've been in the forefront of initiating on the existential risks of what's sometimes referred to as GNR, G for genetics, which is biotechnology, N for nanotechnology, and gray goo, robotics, which is AI.
And I've been accused of being an optimist. And I think you have to be an optimist to be an entrepreneur. If you knew all the problems you were gonna encounter, you'd never start any project. But I've written a lot about the downsides. I remain optimistic. There are specific paradigms, they're not foolproof, that we can follow to keep these technologies safe.
So for example, over 40 years ago, some visionaries recognized the revolutionary potential, both for promise and peril, of biotechnology. Neither the promise nor peril was feasible 40 years ago. But they had a conference at the Asilomar Conference Center in California to develop both professional ethics and strategies to keep biotechnology safe.
And they've been known as the Asilomar Guidelines. They've been refined through successive Asilomar conferences. Much of that's baked into law. And in my opinion, it's worked quite well. We're now, as I mentioned, getting profound benefit. It's a trickle today, it'll be a flood over the next decade. And the number of people who have been harmed, either through intentional or accidental abuse of biotechnology, so far is zero.
Actually, I take that back. There was one boy who died in gene therapy trials about 12 years ago. And there was congressional hearings and they canceled all research for gene therapy for a number of years. You could do an interesting master's thesis and demonstrate that 300,000 people died as a result of that delay, but you can't name them.
They can't go on CNN, so we don't know who they are. But that has to do with the balancing of risk. But in large measure, virtually no one has been hurt by biotechnology. Now, that doesn't mean you can cross it off our list. Okay, we took care of that one because the technology keeps getting more sophisticated.
CRISPR's a great opportunity. There's hundreds of trials of CRISPR technologies to overcome disease, but it could be abused. You can easily describe scenarios, so we have to keep reinventing it. January, we had our first Asilomar Conference on AI ethics. And so I think this is a good paradigm. It's not foolproof.
I think the best way we can assure a democratic future that includes our ideas of liberty is to practice that in the world today 'cause the future world of the singularity, which is a merger of biological and non-biological intelligence, is not gonna come from Mars. I mean, it's gonna emerge from our society today.
So if we practice these ideals today, it's gonna have a higher chance of us practicing them as we get more enhanced with technology. That doesn't sound like a foolproof solution. It isn't, but I think that's the best approach. In terms of technological solutions, I mean, AI is the most daunting.
You can imagine there are technical solutions to biotechnology and nanotechnology. There's really no subroutine you can put in your AI software that will assure that it remains safe. Intelligence is inherently not controllable. If there's some AI that's much smarter than you that's out for your destruction, the best way to deal with that is not to get in that situation in the first place.
If you are in that situation, find some AI that will be on your side. But basically, it's going to, I believe we have been headed through technology to a better reality. I go around the world and people really think things are getting worse. And I think that's 'cause our information about what's wrong with the world is getting exponentially better.
They say, "Oh, this is the most peaceful time "in human history." And people say, "What are you, crazy? "Didn't you hear about the event yesterday and last week?" Well, 100 years ago, there could be a battle that wiped out the next village and you wouldn't even hear about it for months.
I have all these graphs on education and literacy has gone from like 10% to 90% over a century and health, wealth, poverty has declined 95% in Asia over the last 25 years, it's documented by the World Bank. All these trends are very smoothly getting better and everybody thinks things are getting worse.
But you're right, like on violence, that curve could be quite disrupted if there's an existential event. As I say, I'm optimistic, but I think that is something that we need to deal with and a lot of it is not technological, it's dealing with our social, cultural institutions. - So you mentioned also exponential growth of software and ideas, I guess, related to software.
So one of the reasons for which you said that information technology costs is exponential is because of fundamental properties of matter and energy. But in the case of ideas, why would it have to be exponential? - Well, a lot of ideas produce exponential, exponential gains. They don't increase performance linearly.
There was actually a study during the Obama administration by the Scientific Advisory Board on assessing this question, how much gains on 23 classical engineering problems were gained through hardware improvements over the last decade and software improvements. And there's about a thousand to one improvements, it's about doubling every year from hardware.
There was an average of something like 26,000 to one through software improvements, algorithmic improvements. So we do see both, and apparently, if you come up with an advance, it doubles the performance or multiplies it by 10. We see basically exponential growth from each innovation. So, and we certainly see that in deep learning, the architectures are getting better while we also have more data and more computation and more memory to throw at these algorithms.
- Thank you very much. Let's give a very big hand. (audience applauding) Thank you for being here.