Skip to comments.
Singularity Summit At Stanford Explores Future Of 'Superintelligence'
KurzweilAI.net ^
| 4/13/2006
| Staff
Posted on 04/13/2006 7:22:29 AM PDT by Neville72
The Stanford University Symbolic Systems Program and the Singularity Institute for Artificial Intelligence announced today the Singularity Summit at Stanford, a one-day event free to the public, to be held Saturday, May 13, 2006 at Stanford Memorial Auditorium, Stanford, California.
The event will bring together leading futurists and others to examine the implications of the "Singularity" -- a hypothesized creation of superintelligence as technology accelerates over the coming decades -- to address the profound implications of this radical and controversial scenario.
"The Singularity will be a future period during which the pace of technological change will be so rapid, its impact so deep, that human life will be irreversibly transformed," said Ray Kurzweil, keynote speaker and author of the best-selling The Singularity Is Near: When Humans Transcend Biology (Viking, 2005). "Based on models of technology development that I've used to forecast technological change successfully for more than 25 years, I believe computers will pass the Turing Test by 2029, and by the 2040s our civilization will be billions of times more intelligent."
"Some regard the Singularity as a positive event and work to hasten its arrival, while others view it as unlikely, or even dangerous and undesirable," said Todd Davies, associate director of Stanford's Symbolic Systems Program. "The conference will bring together a range of thinkers about AI, nanotechnology, cognitive science, and related areas for a public discussion of these important questions about our future."
Noted speakers at the event will also include cognitive scientist Douglas R. Hofstadter, author of the Pulitzer prize-winning Gödel, Escher, Bach; nanotechnology pioneers K. Eric Drexler and Christine L. Peterson; science-fiction novelist Cory Doctorow; philosopher Nick Bostrom; futurist Max More; Eliezer S. Yudkowsky, research fellow of the Singularity Institute for Artificial Intelligence; Acceleration Studies Foundation president John Smart; PayPal founder and Clarium Capital Management president Peter Thiel; Steve Jurvetson, a Managing Director of Draper Fisher Jurvetson; and Sebastian Thrun, Stanford Artificial Intelligence Laboratory director and Project Lead of the Stanford Racing Team (DARPA Grand Challenge $2 million winner). In addition, author Bill McKibben will participate remotely from Maine via Teleportec, a two-way, life-size 3D display of the speaker.
The event will be moderated by Peter Thiel and Tyler Emerson, executive director of the Singularity Institute for Artificial Intelligence.
Among the issues to be addressed:
Bostrom: Will superintelligence help us reduce or eliminate existential risks, such as the risk that advanced nanotechnology will be used by humans in warfare or terrorism?
Doctorow: Will our technology serve us, or control us?
Drexler: Will productive nanosystems enable the development of more intricate and complex productive systems, creating a feedback loop that drives accelerating change?
Hofstadter: What is the likelihood of our being eclipsed by (or absorbed into) a vast computational network of superminds, in the course of the next few decades?
Kurzweil: Will the Singularity be a soft (gradual) or hard (rapid) take off and how will humans stay in control?
More: Will our emotional, social, psychological, ethical intelligence and self-awareness keep up with our expanding cognitive abilities?
Peterson: How can we safely bring humanity and the biosphere through the Singularity?
Thrun: Where does AI stand in comparison to human-level skills, in light of the recent autonomous robot race, the DARPA Grand Challenge?
Yudkowsky: How can we shape the intelligence explosion for the benefit of humanity?
The Singularity Summit is hosted by the Symbolic Systems Program at Stanford, and co-sponsored by Clarium Capital Management, KurzweilAI.net, MINE, the Singularity Institute for Artificial Intelligence, the Stanford Transhumanist Association, and United Therapeutics.
The free event will be held in Stanford Memorial Auditorium, 551 Serra Mall, Stanford, CA 94305. Seating is limited. Please RSVP. For further information: sss.stanford.edu or 650-353-6063.
TOPICS: Miscellaneous
KEYWORDS: ai; borg; computer; cyborg; evolution; evolutionary; exponentialgrowth; future; futurist; genetics; gnr; humanity; intelligence; knowledge; kurzweil; longevity; luddite; machine; mind; nanotechnology; nonbiological; physics; raykurzweil; robot; robotics; science; singularity; singularityisnear; spike; stanford; superintelligence; technology; thesingularityisnear; transhuman; transhumanism; trend; virtualreality; wearetheborg
Navigation: use the links below to view more comments.
first previous 1-20, 21-40, 41-60, 61-80 ... 121-131 next last
To: CertainInalienableRights
"Considering that most of the people who are supposed intellectual superiors (libs) make some of the most catastrophic decisions in the history of humanity, I'm not sure this singularity is a good idea.
But I'm just a neanderthal conservative.
Maybe instead I should be the first to welcome our singularity overlords..."
I agree with you. Unfortunately, I can see no way to stop this type of thing from happening. Consider this: pretty soon, military systems will start becoming too fast-acting for people to control. It's a little like programmed trading, if you're familiar with that. The computers play games with the stock market that no person can keep up with.
Programmed trading is constricted by law. But consider the military thing. Suppose we suspected that such a system, set up to defend us, was going awry and working against or self-interest. If we pulled the pulled the plug, we would be vulnerable to our enemies. (This scenario presumes we would have enemies on a similar technological level.)
Anyway, if you think about it we will be developing all sorts of dependencies on computers that we can't really reverse or get out of without tremendous costs. Just think of the biggest microprocessors. I'm not up to date on the numbers, but the last I heard they were up to one third of a billion transistors in a single IC chip. Obviously, no person can actually understand this design. Makes you think.
21
posted on
04/13/2006 8:10:31 AM PDT
by
strategofr
(Hillary stole 1000+ secret FBI files on DC movers & shakers, Hillary's Secret War, Poe, p. xiv)
To: NoStaplesPlease
"But AI, I don't buy it. Just because you link up an astonishing amount of processing power does not mean it's going to eventually become self-aware."
Self-awareness is not the most important question. consider, they already have developed a computer/computer program combination that can play pretty much equal with the best human chess player in the world. Chess used to be considered one of the highest measures of human intelligence. In a short while (if they choose to do it) they can make a computer that can crush any person in chess.
There simply is no limit to this process of development, unfortunately.
22
posted on
04/13/2006 8:13:39 AM PDT
by
strategofr
(Hillary stole 1000+ secret FBI files on DC movers & shakers, Hillary's Secret War, Poe, p. xiv)
To: DoctorMichael
""Grog no like Superintelligence"."
I'm with you, Grog. Any room in the cave, there?
23
posted on
04/13/2006 8:14:36 AM PDT
by
strategofr
(Hillary stole 1000+ secret FBI files on DC movers & shakers, Hillary's Secret War, Poe, p. xiv)
To: strategofr
I'm with you, Grog. Any room in the cave, there? At least he will be alive. There's no assurance SI (Singularity Intelligence) will tolerate humans and their thought patterns.
To: strategofr
Unfortunately, I can see no way to stop this type of thing from happening. Maybe SI should be quarentined in space, or a separate human colony in space be established beforehand that can defend itself and destroy a rogue SI, should things go terribly wrong.
To: NoStaplesPlease
Ultimately I don't know how you test for true self-awareness compared simply to well-mimicked self-awareness. A very complex computer could very persuasively imitate human intelligence, sure. But actually think for itself? Good point. No matter how convincing the test there will always be people who refuse to believe the AI is self aware. I wonder if this would lead to the next step in civil rights.
Regardless of how intelligence begins -- whether spiritual or physical -- it seems to me there must be a spark, a jump-start, a something-else beyond computing ability. We're not the sum of our brain's computing power. There's something mysterious going on in there, and until we can describe that mysteriousness, we're not going to be able to create it in machines.
If a computer can become self aware, does it have the ability to believe that God is self evident? If he does, do you think that it might then have a soul?
I've gotta run now (work and all that) but I'll check back tonight. Thanks.
26
posted on
04/13/2006 8:31:51 AM PDT
by
SunTzuWu
(Hans Delbruck - Scientist and Saint.)
To: Neville72
It's all crap. It's a bunch of people who read too much scifi and wish it were real. Not that I have anything against scifi, it's may favorite literature, but it's just entertainment.
When I was in school, I had a professor that was working a lot with robotics and AI. It wasn't a class in our computer science program, but he did talk about it a lot.
The problem you run into is that the self-aware human mind exhibits some qualities, some of which are difficult to put a finger on, that a solid-state electronic computer is physically incapapble of reproducing, no matter how complicated it is.
A computer program can be theoretically modeled with something called a state-transition diagram. This diagram represents every single possible state the computer could be in, and how it transitions from state to state. As an academic exercise, you might design a state-transition diagram that causes a computer to go into an accept state when a certain string is input. This could be drawn on one page. The diagram for something like Windows XP, however, would be so large that I'm nearly certain no one has ever bothered making one. However, if they did, what they could do is describe, to the minute detail, every single possible thing Windows could ever do. And anything not in that diagram is something the program could not do, ever, under any circumstances.
The human brain does not work this way, unless we truly are the sum of our parts. Human beings come with some basic 'software' installed. We call them instincts. Unlike a computer, which has no choice but to obey its programming, we can ignore our own instincts if we choose to.
It's almost an issue of free will. Computers do the things they do because they literally have no choice. They can't choose what to do or what not to do anymore than the sun could choose whether or not to quit shining or the snow could choose whether or not to be cold. Human beings, however, have the ability to do this, which is almost paradoxical; the ability to choose anything you want suggests that true randomness exists and the universe is non-deterministic, or at least that the universe allows non-determinism. Computers, however, are remarkably deterministic. Even a random number generator in a computer isn't really random, it just generates a large enough set of numbers to be good enough in most cases. Feed it the same random seed value and you'll get exactly the same sequence of not-so-random numbers. If the universe, however, is deterministic and not non-deterministic, then human beings really don't have free will and any thought that you did is simply a lie, or rather you had that thought because you were programmed to and had no choice in the matter. As for me, I don't believe that. I think we do have free will, a precious gift granted to mankind by no less than God Himself. Anyway, that's my personal opinion. Your mileage will probably vary.
As long as computers are built with solid state components, I think it's physically impossible for them to have intelligence, short of divine intervention by God Himself (a possibility that I don't count out, but that's another thread). Computers that function on a non-deterministic principle have the potential to have intelligence or self-awareness. The only two ways to *maybe* accomplish this that I can see is to either use wetware or quantum computing.
Quantum physics is highly chaotic, and any computer based on it would have potential to be non-deterministic.
Wetware solutions would include using cloned brains instead of CPUs and hard drives to create a self-aware computer. However, once you do that, I don't really think it qualifies as a computer anymore.
Anyway, these people are a little crazy, in my opinion. Creating true AI is not as simple as they make it sound, and it may not be desireable either. I know we've seen plenty of scifi like the Matrix that deals with murderious AIs. We don't know for certain that any computer with intelligence wouldn't turn out to be a nice guy with a sense of civic responsbility that loves kids. OTOH, we don't know that it wouldn't go psycho on us either.
Honestly, we know so little about natural intelligence that we really can't even define it properly, much less manufacture it. These people are ahead of themselves.
27
posted on
04/13/2006 8:36:30 AM PDT
by
JamesP81
(Socialism is based on how things should be. Capitalism is based on how things are, and deals with it)
To: NoStaplesPlease
See my post #27 for what I think of that.
28
posted on
04/13/2006 8:39:28 AM PDT
by
JamesP81
(Socialism is based on how things should be. Capitalism is based on how things are, and deals with it)
To: Neville72
Yudkowsky: How can we shape the intelligence explosion for the benefit of humanity? What if the answer is: "We can't".
To: strategofr
Self-awareness is not the most important question. consider, they already have developed a computer/computer program combination that can play pretty much equal with the best human chess player in the world. Chess used to be considered one of the highest measures of human intelligence.
It's my opinion that that was an invalid experiment. Gary Kasparov is a chess player. His environment is one where he sits down at the chess board across from another player and attempts to win.
Kasparov went into the game against big blue with same mindset he had always had. Play against another chess player.
He lost because that wasn't the game he went into. He went in playing against not a computer, but a programmer. Superficially, it seemed to be the same contest he was familiar with, but in truth it was totally different. It really wasn't a chess game anymore, but something else with chess as the window dressing.
The programmer better understood the rules and the environment of the contest than Kasparov did. If Kasparov had gone in with the mindset of defeating a programmer's toy at chess, there's a much higher chance he'd have won.
30
posted on
04/13/2006 8:44:58 AM PDT
by
JamesP81
(Socialism is based on how things should be. Capitalism is based on how things are, and deals with it)
To: Fitzcarraldo
Thanks...you mean Hubert Dreyfus, the professor at UCB? That was a good link. Modern AI embraces the idea of handling the gazillions of special cases of real life, rather than abstracting them to symbols and rules. Fluid representations dominate.
31
posted on
04/13/2006 8:45:04 AM PDT
by
no-s
To: JamesP81
Anyway, these people are a little crazy, in my opinion. With them, it's an all or nothing situation. That's extremely dangerous, given that they don't account for the unfavorable outcomes at all, assuming everything will be rosy if THEY design the "seed" SI.
To: JamesP81
The programmer better understood the rules and the environment of the contest than Kasparov did. If Kasparov had gone in with the mindset of defeating a programmer's toy at chess, there's a much higher chance he'd have won. That's the problem. The machine could "win" to the detriment of the human race. At some point even the programmer of SI will not be able to keep up.
To: no-s
Modern AI embraces the idea of handling the gazillions of special cases of real life, rather than abstracting them to symbols and rules. SI capable of exploiting nanotechnology will be unstoppable. Its thought processes will be alien to humans. You've stated it very well.
To: SunTzuWu
If we succeed in creating an AI, will that change your views on religion or make you an atheist?
For me, personally, no. It would be no different that a test tube baby: it's still a person, and destroying it would still constitute murder before God. Any truly artifically intelligent computer would, IMO, be legally entitled to the same constitutional rights as anyone else.
Here's an interesting thought: such a being could, in a very real sense, be considered an alien (not the illegal kind, the little green men kind) because it's an intelligent being that's not human (unless we base it on a cloned human brain, a la wetware solution I talked about in #27).
35
posted on
04/13/2006 8:52:44 AM PDT
by
JamesP81
(Socialism is based on how things should be. Capitalism is based on how things are, and deals with it)
To: Fitzcarraldo
What if the answer is: "We can't".
Then let's hope it isn't created. As a computer scientist, I concede the possibility someone might do it. OTOH, even if they did, I think we would still win the ensuing war, eventually.
There are two things in this universe I have faith in. One is God's mercy. The other is the human ability to inflict devastation. Basically, I don't think an AI would have the sheer d@mned bloodthirsty meanness necessary to kill us all before we took it down.
36
posted on
04/13/2006 8:56:44 AM PDT
by
JamesP81
(Socialism is based on how things should be. Capitalism is based on how things are, and deals with it)
To: NoStaplesPlease
But AI, I don't buy it. Just because you link up an astonishing amount of processing power does not mean it's going to eventually become self-aware.
Some very smart people seem to think that's how it works, as if once there's enough power, it just happens. Maybe if you're an atheist, you think it does.
Ultimately I don't know how you test for true self-awareness compared simply to well-mimicked self-awareness.
A very complex computer could very persuasively imitate human intelligence, sure.
But actually think for itself? I believe this would have to be an illusion.
Regardless of how intelligence begins -- whether spiritual or physical -- it seems to me there must be a spark, a jump-start, a something-else beyond computing ability. We're not the sum of our brain's computing power.
There's something mysterious going on in there, and until we can describe that mysteriousness, we're not going to be able to create it in machines.
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Mirror Test
One benchmark for "self-awareness" in animals and people (and now robots as well) is whether they will perform self-directed actions when looking in a mirror. When a mark is placed on the forehead of a child, they will only begin to inspect it on their own forehead at the age of 3 or 4. Adult bottlenose dolphins perform similarly in equivalent tests designed for underwater use.
According to this discovery news article, Junichi Takeno and a team of researchers at Meiji University in Japan have observed similar behavior in a robot with a hierarchical neural network.
Developing Intelligence: Imitation vs Self-awareness: The Mirror Test
Address:http://develintel.blogspot.com/2005/12/imitation-vs-self-awareness-mirror.html
37
posted on
04/13/2006 8:57:58 AM PDT
by
tpaine
To: JamesP81
because it's an intelligent being that's not human (unless we base it on a cloned human brain, I don't anticipate any restrictions or ethics or primary design principles will be implemented save those that expand the SI's capabilities as quickly as possible. We've seen this happen before in human technological development, save for the restrictions the Amish place on introduction of new techonology to their culture. SI is to the 21st century as nuclear weapons was to the 20th. The only difference is that nuclear weapons are controlled by humans. SI will not be.
To: tpaine; NoStaplesPlease; All
I think our worries would be much better spent worrying about 'dumb' computer systems under the control of wacked out humans, a la President Amhadi-nejad in Iran.
Another thing to consider: they're working on creating neural interfaces. If that's ever perfected, a human linked to a computer would have all the advantages of being human, plus the reaction time and computing capability of a modern computer. I submit to you that no AI would ever be superior to that. In fact, it would probably be a few grades inferior.
39
posted on
04/13/2006 9:01:27 AM PDT
by
JamesP81
(Socialism is based on how things should be. Capitalism is based on how things are, and deals with it)
To: Fitzcarraldo
There are concerns about SI, but I'm not yet convinced we will have the technology to do it in this century. I mentioned possibly using wetware and quantum computers to do it, but that's just educated guesswork. It *might* work, and it may not.
Humans also have a blessed disinclination to carry anything to it's ultimate apocalyptic conclusion. Hopefully, that trend will hold if it turns out that it is possible.
40
posted on
04/13/2006 9:04:19 AM PDT
by
JamesP81
(Socialism is based on how things should be. Capitalism is based on how things are, and deals with it)
Navigation: use the links below to view more comments.
first previous 1-20, 21-40, 41-60, 61-80 ... 121-131 next last
Disclaimer:
Opinions posted on Free Republic are those of the individual
posters and do not necessarily represent the opinion of Free Republic or its
management. All materials posted herein are protected by copyright law and the
exemption for fair use of copyrighted works.
FreeRepublic.com is powered by software copyright 2000-2008 John Robinson