While reading "The Intelligent Universe" by james gardener..................
there was a mention of the technological singularity....and i was awed by its beauty and yes to be honest somewhat concerned as well...............
John von Neumann was the first person to give a thought
about the implications of a looming technological singularity:
One conversation centered on the ever accelerating progress of technology and changes in the mode of human life, which gives the appearance of
approaching some essential singularity in the history of the race beyond which human affairs, as we know them, could not continue.
And then the statiscian I.J.Good remarked
Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the
design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines; there would then
unquestionably be an “intelligence explosion,” and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is
the last invention that man need ever make.
At a conference the famous context made about technological singularity...
Within thirty years, we will have the technological means to create superhuman intelligence. Shortly after, the human era will be ended. Is such
progress avoidable? If not to be avoided, can events be guided so that we may survive? These questions are investigated. Some possible answers (and
some further dangers) are presented.
What are the consequences of this event? When greater-than-
human intelligence drives progress, that progress will be much
more rapid. In fact, there seems no reason why progress itself
would not involve the creation of still more intelligent
entities—on a still-shorter time scale. The best analogy that I
see is with the evolutionary past: Animals can adapt to problems
and make inventions, but often no faster than natural selection
can do its work—the world acts as its own simulator in the
case of natural selection. We humans have the ability to
internalize the world and conduct “what ifs” in our heads; we
can solve many problems thousands of times faster than natural
selection. Now by creating the means to execute those
simulations at much higher speeds, we are entering a regime
as radically different from our own human past as we humans
are from the lower animals.
The lessons of our evolutionary past were, in Vinge’s view, not exactly
From the human point of view this change will be a throwing
away of all previous rules, perhaps in the blink of an eye, an
exponential runaway beyond any hope of control.
Developments that before were thought might only happen
in “a million years” (if ever) will likely happen in the next
century. [One commentator] paints a picture of the major
changes happening in a matter of hours.... [The most
disturbing consequence of the technological singularity is that
any hyper-intelligent machine] would not be humankind’s
“tool”—any more than humans are the tools of rabbits or
robins or chimpanzees
srry didnt know where to post this so I put it here.................feel free to move it under the suitable forum