Jump to content

Talk:Technological singularity

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by M@ (talk | contribs) at 14:20, 27 October 2011. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Former good articleTechnological singularity was one of the Social sciences and society good articles, but it has been removed from the list. There are suggestions below for improving the article to meet the good article criteria. Once these issues have been addressed, the article can be renominated. Editors may also seek a reassessment of the decision if they believe there was a mistake.
Article milestones
DateProcessResult
August 7, 2005Peer reviewReviewed
July 19, 2007Good article nomineeListed
July 7, 2008Good article reassessmentDelisted
Current status: Delisted good article

Bots

Should we be allowing bots to edit this article? Think about it.

Is this a religion?

I saw a book describing the Singularity as a new religion:

"The idea that information is alive in its own right is a metaphysical claim made by people who hope to become immortal by being uploaded into a computer someday. It is part of what should be understood as a new religion. That might sound like an extreme claim, but go visit any computer science lab and you’ll find books about "the Singularity," which is the supposed future event when the blessed uploading is to take place. A weird cult in the world of technology has done damage to culture at large."

http://www.amazon.com/You-Are-Not-Gadget-Manifesto/dp/0307389979/ref=pd_sim_b_6

Should this be noted in the article? —Preceding unsigned comment added by 24.99.60.219 (talk) 01:21, 10 May 2011 (UTC)[reply]

Well, as already noted, the singularity has been described as "the Rapture for nerds".
—WWoods (talk) 20:09, 10 May 2011 (UTC)[reply]
AI heal Thyself
Urgen (talk) 04:20, 6 June 2011 (UTC)[reply]


I think "religion" is the wrong word. The Singularity concept certainly represents a metaphysical perspective and a motivation for computer evolution that goes beyond mere short-term utility and profit. The motivation is driven rather by a much larger and longer term vision of the human condition and the profound transformations that will effect our life on this planet and in the universe. — Preceding unsigned comment added by 69.86.25.123 (talk) 17:46, 23 October 2011 (UTC)[reply]

Just a Thought

Let's suppose that at some point, AI could truly simulate a human's thought. AI, then, would possess the qualities of human thinking, specifically decision-making and the ability to weigh both outcomes (the "good" one and the "bad"/evil one). If we explain crime in human behavior as the decision to indulge in the criminal behavior with consideration to the possibility of receiving punishment and impunitively not receiving punishment, then the existance of super intellegence seems to cause logical issues. If a superintellegence knew that it could elude punishment from both humans and other superintellegent entities, then what would restrain this superintellegent entity from being infinitely criminal (when there exists corresponding opportunities to make gains through such conduct)? 74.195.236.12 (talk) 00:49, 25 May 2011 (UTC)[reply]

Avoiding criminal behavior is not necessarily solely due to fear of punishment. I don't steal because it's contrary to social contract (Golden Rule, etc), because I realize that society could not survive if every person was a thief. I voluntarily sacrifice some rights, and modify some behaviors, to be consistent with a productive, functioning member of a society - precisely because I know I derive benefit from said society. My behavior is not dictated by fear of punishment, but rather from not contributing to the loss of the society that I currently benefit from. — Preceding unsigned comment added by 76.203.229.161 (talk) 03:57, 7 July 2011 (UTC)[reply]

The whole topic is a fallacy

What's all this talk about artificial intelligence taking over the world from human beings? Since when does intelligence hold power on Earth? Stupidity has, and will always be, the most powerful force on Earth, and intelligent beings and institutions have always been subservient to it or only able, at most, to garner some fringe benefits. Enough said.

Yet we are not ruled by chimps. -- cheers, Michael C. Price talk 17:16, 23 June 2011 (UTC)[reply]

Who says human stupidity will not gift power to AI? It's not inconceivable that humanity will place increasing reliance on AI to manage life and eventually hand over the keys to everything?

Updated term

Updated the term to this:

Technological singularity refers to the point in time where machines have become self-educating and are doing it so fast that humans are no longer able to keep up with hem.[1] Since the capabilities of such a greater-than human intelligence are too difficult for an unaided human mind to comprehend, a technological singularity is seen as an intellectual event horizon, beyond which the future becomes (too) difficult to understand or predict.

Also added link to Eureqa.

Can it also be mentioned that the technological singularity (atleast on the determining of natural laws) has allready been reached. According to Hod Lipson, due to Eureqa, we are allready in the post-technological singularity era.

91.182.242.40 (talk) 15:04, 1 September 2011 (UTC)[reply]

Please add this proof that the “singularity” model is bullshit:

Please go to Dr. Albert A. Bartlett's presentation on “Arithmetic, Population, and Energy” (part 5) and jump to 4:48 (I recommend watching all the parts up until that time before, so you can understand the implications.), for proper proof of why a singularity can and will never happen, and the whole thing is Kurzweil’s typical pseudo-scientific nonsense.
94.221.17.98 (talk) 23:38, 24 October 2011 (UTC)[reply]

  1. ^ nwtmagazine, juli/august 2011