Singularity (technology): Difference between revisions
Jump to navigation
Jump to search
imported>Mike Johnson (Started stub; first stab at outlining structure of article) |
imported>Mike Johnson m (Added CZ Live designation) |
||
Line 7: | Line 7: | ||
# [[Vernor Vinge]]'s singularity refers to the concept of rapidly accelerating change creating an [[event horizon]]-like barrier to social prediction; | # [[Vernor Vinge]]'s singularity refers to the concept of rapidly accelerating change creating an [[event horizon]]-like barrier to social prediction; | ||
# [[I.J. Good]]'s singularity refers to the concept that intelligent agents who are able to improve their own intelligence will cause an [[intelligence explosion]]. | # [[I.J. Good]]'s singularity refers to the concept that intelligent agents who are able to improve their own intelligence will cause an [[intelligence explosion]]. | ||
[[Category:CZ Live]] |
Revision as of 11:22, 26 September 2007
The term "technological singularity" is used in futurist circles to refer to the phenomenon and resulting effects of reaching a critical threshold of positive-feedback technological change.
Three Models
Eliezer Yudkowsky, co-founder of the Singularity Institute, has suggested that the term 'technological singularity' holds three distinct concepts[1]:
- Ray Kurzweil's singularity refers to the concept of rapidly accelerating change driven by technology;
- Vernor Vinge's singularity refers to the concept of rapidly accelerating change creating an event horizon-like barrier to social prediction;
- I.J. Good's singularity refers to the concept that intelligent agents who are able to improve their own intelligence will cause an intelligence explosion.
- ↑ [Introducing the "Singularity": Three Major Schools of Thought. Singularity Summit 2007.