Blog Archive

Friday, February 10, 2023

02-10-2023-1336 - intelligence amplification distributed cognition extended mind thesis technological singularity intelligence explosion existential risk from artificial general intelligence



https://en.wikipedia.org/wiki/Intelligence_amplification

https://en.wikipedia.org/wiki/Distributed_cognition

https://en.wikipedia.org/wiki/Extended_mind_thesis

https://en.wikipedia.org/wiki/Technological_singularity#Intelligence_explosion

https://en.wikipedia.org/wiki/Existential_risk_from_artificial_general_intelligence



In Max Tegmark's 2017 book Life 3.0, a corporation's "Omega team" creates an extremely powerful AI able to moderately improve its own source code in a number of areas, but after a certain point the team chooses to publicly downplay the AI's ability, in order to avoid regulation or confiscation of the project. For safety, the team keeps the AI in a box where it is mostly unable to communicate with the outside world, and tasks it to flood the market through shell companies, first with Amazon Mechanical Turk tasks and then with producing animated films and TV shows. Later, other shell companies make blockbuster biotech drugs and other inventions, investing profits back into the AI. The team next tasks the AI with astroturfing an army of pseudonymous citizen journalists and commentators, in order to gain political influence to use "for the greater good" to prevent wars. The team faces risks that the AI could try to escape via inserting "backdoors" in the systems it designs, via hidden messages in its produced content, or via using its growing understanding of human behavior to persuade someone into letting it free. The team also faces risks that its decision to box the project will delay the project long enough for another project to overtake it.[37][38]

https://en.wikipedia.org/wiki/Existential_risk_from_artificial_general_intelligence

Bostrom, Russell, and others argue that smarter-than-human decision-making systems could arrive at more unexpected and extreme solutions to assigned tasks, and could modify themselves or their environment in ways that compromise safety requirements.[5][7]

https://en.wikipedia.org/wiki/Existential_risk_from_artificial_general_intelligence
Paperclip maximizer

The paperclip maximizer is a thought experiment described by Swedish philosopher Nick Bostrom in 2003. It illustrates the existential risk that an artificial general intelligence may pose to human beings when programmed to pursue even seemingly harmless goals, and the necessity of incorporating machine ethics into artificial intelligence design. The scenario describes an advanced artificial intelligence tasked with manufacturing paperclips. If such a machine were not programmed to value human life, then given enough power over its environment, it would try to turn all matter in the universe, including human beings, into either paperclips or machines which manufacture paperclips.[4]


Suppose we have an AI whose only goal is to make as many paper clips as possible. The AI will realize quickly that it would be much better if there were no humans because humans might decide to switch it off. Because if humans do so, there would be fewer paper clips. Also, human bodies contain a lot of atoms that could be made into paper clips. The future that the AI would be trying to gear towards would be one in which there were a lot of paper clips but no humans.
— Nick Bostrom, as quoted in Miles, Kathleen (2014-08-22). "Artificial Intelligence May Doom The Human Race Within A Century, Oxford Professor Says". Huffington Post.[5]

Bostrom has emphasised that he does not believe the paperclip maximiser scenario per se will actually occur; rather, his intention is to illustrate the dangers of creating superintelligent machines without knowing how to safely program them to eliminate existential risk to human beings.[6] The paperclip maximizer example illustrates the broad problem of managing powerful systems that lack human values.[7]
Delusion and survival

The "delusion box" thought experiment argues that certain reinforcement learning agents prefer to distort their own input channels to appear to receive high reward; such a "wireheaded" agent abandons any attempt to optimize the objective in the external world that the reward signal was intended to encourage.[8] The thought experiment involves AIXI, a theoretical[a] and indestructible AI that, by definition, will always find and execute the ideal strategy that maximizes its given explicit mathematical objective function.[b] A reinforcement-learning[c] version of AIXI, if equipped with a delusion box[d] that allows it to "wirehead" its own inputs, will eventually wirehead itself in order to guarantee itself the maximum reward possible, and will lose any further desire to continue to engage with the external world. As a variant thought experiment, if the wireheadeded AI is destructable, the AI will engage with the external world for the sole purpose of ensuring its own survival; due to its wireheading, it will be indifferent to any other consequences or facts about the external world except those relevant to maximizing the probability of its own survival.[10] In one sense AIXI has maximal intelligence across all possible reward functions, as measured by its ability to accomplish its explicit goals; AIXI is nevertheless uninterested in taking into account what the intentions were of the human programmer.[11] This model of a machine that, despite being otherwise superintelligent, appears to simultaneously be stupid (that is, to lack "common sense"), strikes some people as paradoxical.[12]
Basic AI drives

Steve Omohundro has itemized several convergent instrumental goals, including self-preservation or self-protection, utility function or goal-content integrity, self-improvement, and resource acquisition. He refers to these as the "basic AI drives". A "drive" here denotes a "tendency which will be present unless specifically counteracted";[13] this is different from the psychological term "drive", denoting an excitatory state produced by a homeostatic disturbance.[14] A tendency for a person to fill out income tax forms every year is a "drive" in Omohundro's sense, but not in the psychological sense.[15] Daniel Dewey of the Machine Intelligence Research Institute argues that even an initially introverted self-rewarding AGI may continue to acquire free energy, space, time, and freedom from interference to ensure that it will not be stopped from self-rewarding.[16]
Goal-content integrity

In humans, maintenance of final goals can be explained with a thought experiment. Suppose a man named "Gandhi" has a pill that, if he took it, would cause him to want to kill people. This Gandhi is currently a pacifist: one of his explicit final goals is to never kill anyone. Gandhi is likely to refuse to take the pill, because Gandhi knows that if in the future he wants to kill people, he is likely to actually kill people, and thus the goal of "not killing people" would not be satisfied.[17]

However, in other cases, people seem happy to let their final values drift. Humans are complicated, and their goals can be inconsistent or unknown, even to themselves.[18]
In artificial intelligence

In 2009, Jürgen Schmidhuber concluded, in a setting where agents search for proofs about possible self-modifications, "that any rewrites of the utility function can happen only if the Gödel machine first can prove that the rewrite is useful according to the present utility function."[19][20] An analysis by Bill Hibbard of a different scenario is similarly consistent with maintenance of goal content integrity.[20] Hibbard also argues that in a utility maximizing framework the only goal is maximizing expected utility, so that instrumental goals should be called unintended instrumental actions.[21]
Resource acquisition

Many instrumental goals, such as resource acquisition, are valuable to an agent because they increase its freedom of action.[22]

For almost any open-ended, non-trivial reward function (or set of goals), possessing more resources (such as equipment, raw materials, or energy) can enable the AI to find a more "optimal" solution. Resources can benefit some AIs directly, through being able to create more of whatever stuff its reward function values: "The AI neither hates you, nor loves you, but you are made out of atoms that it can use for something else."[23][24] In addition, almost all AIs can benefit from having more resources to spend on other instrumental goals, such as self-preservation.[24]
Cognitive enhancement

"If the agent's final goals are fairly unbounded and the agent is in a position to become the first superintelligence and thereby obtain a decisive strategic advantage, [...] according to its preferences. At least in this special case, a rational intelligent agent would place a very high instrumental value on cognitive enhancement"[25]
Technological perfection

Many instrumental goals, such as [...] technological advancement, are valuable to an agent because they increase its freedom of action.[22]
Self-preservation

Many instrumental goals, such as self-preservation, are valuable to an agent because they increase its freedom of action.[22]
Instrumental convergence thesis

The instrumental convergence thesis, as outlined by philosopher Nick Bostrom, states:


Several instrumental values can be identified which are convergent in the sense that their attainment would increase the chances of the agent's goal being realized for a wide range of final goals and a wide range of situations, implying that these instrumental values are likely to be pursued by a broad spectrum of situated intelligent agents.

The instrumental convergence thesis applies only to instrumental goals; intelligent agents may have a wide variety of possible final goals.[3] Note that by Bostrom's orthogonality thesis,[3] final goals of highly intelligent agents may be well-bounded in space, time, and resources; well-bounded ultimate goals do not, in general, engender unbounded instrumental goals.[26]
Impact

Agents can acquire resources by trade or by conquest. A rational agent will, by definition, choose whatever option will maximize its implicit utility function; therefore a rational agent will trade for a subset of another agent's resources only if outright seizing the resources is too risky or costly (compared with the gains from taking all the resources), or if some other element in its utility function bars it from the seizure. In the case of a powerful, self-interested, rational superintelligence interacting with a lesser intelligence, peaceful trade (rather than unilateral seizure) seems unnecessary and suboptimal, and therefore unlikely.[22]

Some observers, such as Skype's Jaan Tallinn and physicist Max Tegmark, believe that "basic AI drives", and other unintended consequences of superintelligent AI programmed by well-meaning programmers, could pose a significant threat to human survival, especially if an "intelligence explosion" abruptly occurs due to recursive self-improvement. Since nobody knows how to predict when superintelligence will arrive, such observers call for research into friendly artificial intelligence as a possible way to mitigate existential risk from artificial general intelligence.[27]
See alsoAI control problem
AI takeovers in popular culture Universal Paperclips, an incremental game featuring a paperclip maximizer
Friendly artificial intelligence
Instrumental and intrinsic value
The Sorcerer's Apprentice

 
https://en.wikipedia.org/wiki/Instrumental_convergence#Paperclip_maximizer

https://en.wikipedia.org/wiki/Global_catastrophic_risk#Artificial_intelligence

https://en.wikipedia.org/wiki/Accelerating_change

https://en.wikipedia.org/wiki/Cashless_society

His subsequent Hugo award-winning novel A Fire Upon the Deep (1992) starts with an imaginative description of the evolution of a superintelligence passing through exponentially accelerating developmental stages ending in a transcendent, almost omnipotent power unfathomable by mere humans. His already mentioned influential 1993 paper on the technological singularity compactly summarizes the basic ideas. 

An updated version of Moore's Law over 120 years (based on Kurzweil's graph). The seven most recent data points are all Nvidia GPUs.

 

https://en.wikipedia.org/wiki/Accelerating_change

https://en.wikipedia.org/wiki/Simulated_reality

https://en.wikipedia.org/wiki/Transhumanism

https://en.wikipedia.org/wiki/Technological_utopianism

History

Precursors of transhumanism

According to Nick Bostrom, transcendentalist impulses have been expressed at least as far back as the quest for immortality in the Epic of Gilgamesh, as well as in historical quests for the Fountain of Youth, the Elixir of Life, and other efforts to stave off aging and death.[2]

In his Divine Comedy, Dante coined the word trasumanar meaning "to transcend human nature, to pass beyond human nature" in the first canto of Paradiso.[9][10][11][12]

One of the early precursors to transhumanist ideas is Discourse on Method (1637) by René Descartes. In the Discourse, Descartes envisioned a new kind of medicine that could grant both physical immortality and stronger minds.[13]

In his first edition of Political Justice (1793), William Godwin included arguments favoring the possibility of "earthly immortality" (what would now be called physical immortality). Godwin explored the themes of life extension and immortality in his gothic novel St. Leon, which became popular (and notorious) at the time of its publication in 1799, but is now mostly forgotten. St. Leon may have provided inspiration for his daughter Mary Shelley's novel Frankenstein.[14]

There is debate about whether the philosophy of Friedrich Nietzsche can be considered an influence on transhumanism, despite its exaltation of the "Ãœbermensch" (overman or superman), due to its emphasis on self-actualization rather than technological transformation.[2][15][16][17] The transhumanist philosophies of Max More and Stefan Lorenz Sorgner have been influenced strongly by Nietzschean thinking.[15] By way of contrast, The Transhumanist Declaration[18] "...advocates the well-being of all sentience (whether in artificial intellects, humans, posthumans, or non-human animals)".

The late 19th to early 20th century movement known as Russian cosmism, by Russian philosopher N. F. Fyodorov, is noted for anticipating transhumanist ideas.[19]


https://en.wikipedia.org/wiki/Transhumanism

https://en.wikipedia.org/wiki/Geomagnetic_storm

https://en.wikipedia.org/wiki/Coronal_hole

https://en.wikipedia.org/wiki/Heliospheric_current_sheet

https://en.wikipedia.org/wiki/Geomagnetic_storm

https://en.wikipedia.org/wiki/Interplanetary_magnetic_field

https://en.wikipedia.org/wiki/Plasma_(physics)

https://en.wikipedia.org/wiki/Magnetic_reconnection

https://en.wikipedia.org/wiki/Particle_acceleration

https://en.wikipedia.org/wiki/Instrumental_convergence#Goal-content_integrity

https://en.wikipedia.org/wiki/Instrumental_and_intrinsic_value

https://en.wikipedia.org/wiki/Instrumental_and_value-rational_action

https://en.wikipedia.org/wiki/Intrinsic_value_(ethics)

https://en.wikipedia.org/wiki/Aesthetics

https://en.wikipedia.org/wiki/Deontology

https://en.wikipedia.org/wiki/Intrinsic_and_extrinsic_properties_(philosophy)

https://en.wikipedia.org/wiki/Ontology

https://en.wikipedia.org/wiki/Utilitarianism

https://en.wikipedia.org/wiki/Morality

https://en.wikipedia.org/wiki/Synonym

https://en.wikipedia.org/wiki/Semantic_field

https://en.wikipedia.org/wiki/Computational_semiotics

https://en.wikipedia.org/wiki/Logic

https://en.wikipedia.org/wiki/Linguistics

https://en.wikipedia.org/wiki/Syntax

https://en.wikipedia.org/wiki/Morpheme

https://en.wikipedia.org/wiki/Latin

https://en.wikipedia.org/wiki/Classical_language

https://en.wikipedia.org/wiki/Modern_language

https://en.wikipedia.org/wiki/English_language

https://en.wikipedia.org/wiki/Multilingualism

https://en.wikipedia.org/wiki/Monolingualism

https://en.wikipedia.org/wiki/Language_policy

https://en.wikipedia.org/wiki/Grimm%27s_law

https://en.wikipedia.org/wiki/Language_tax

https://en.wikipedia.org/wiki/Anglosphere

https://en.wikipedia.org/wiki/English-speaking_world

https://en.wikipedia.org/wiki/De_facto

https://en.wikipedia.org/wiki/Technology

https://en.wikipedia.org/wiki/Pleistocene

Milankovitch cycles

Glaciation in the Pleistocene was a series of glacials and interglacials, stadials and interstadials, mirroring periodic changes in climate. The main factor at work in climate cycling is now believed to be Milankovitch cycles. These are periodic variations in regional and planetary solar radiation reaching the Earth caused by several repeating changes in the Earth's motion.

Milankovitch cycles cannot be the sole factor responsible for the variations in climate since they explain neither the long term cooling trend over the Plio-Pleistocene, nor the millennial variations in the Greenland Ice Cores. Milankovitch pacing seems to best explain glaciation events with periodicity of 100,000, 40,000, and 20,000 years. Such a pattern seems to fit the information on climate change found in oxygen isotope cores. 

Oxygen isotope ratio cycles

In oxygen isotope ratio analysis, variations in the ratio of 18
O
to 16
O
(two isotopes of oxygen) by mass (measured by a mass spectrometer) present in the calcite of oceanic core samples is used as a diagnostic of ancient ocean temperature change and therefore of climate change. Cold oceans are richer in 18
O
, which is included in the tests of the microorganisms (foraminifera) contributing the calcite.

A more recent version of the sampling process makes use of modern glacial ice cores. Although less rich in 18
O
than sea water, the snow that fell on the glacier year by year nevertheless contained 18
O
and 16
O
in a ratio that depended on the mean annual temperature.

Temperature and climate change are cyclical when plotted on a graph of temperature versus time. Temperature coordinates are given in the form of a deviation from today's annual mean temperature, taken as zero. This sort of graph is based on another of isotope ratio versus time. Ratios are converted to a percentage difference from the ratio found in standard mean ocean water (SMOW).

The graph in either form appears as a waveform with overtones. One half of a period is a Marine isotopic stage (MIS). It indicates a glacial (below zero) or an interglacial (above zero). Overtones are stadials or interstadials.

According to this evidence, Earth experienced 102 MIS stages beginning at about 2.588 Ma BP in the Early Pleistocene Gelasian. Early Pleistocene stages were shallow and frequent. The latest were the most intense and most widely spaced.

By convention, stages are numbered from the Holocene, which is MIS1. Glacials receive an even number; interglacials, odd. The first major glacial was MIS2-4 at about 85–11 ka BP. The largest glacials were 2, 6, 12, and 16; the warmest interglacials, 1, 5, 9 and 11. For matching of MIS numbers to named stages, see under the articles for those names.

 

https://en.wikipedia.org/wiki/Pleistocene

 

No comments:

Post a Comment