What is the alignment problem in AI?
A failure of value alignment means that humans give artificial intelligence a goal to achieve and it has very negative side effects for individuals or society in general. Nonetheless, the original goal given to machines to solve may be perfectly innocent and rational. An example are the algorithms for the optimisation of click-through rates on social media. The goal is perfectly innocent: Select content that allows users to click on links so that people stay social media platforms longer. The negative side effect is that more and more extreme content engages people, enables click-through and results in radicalisation.An AI can also try to achieve goals that are perfectly innocent to begin with. For instance, a machine is giving the goal to get some coffee. The machine will immediately realise that it cannot achieve the goal if it is destroyed or switched off. Hence, the machine will create a subgoal to ensure that it is operational at all times. This may include self-modification so that it cannot be switched off by a human user. It is possible to generalise this problem immediately to realise that achieving any goal requires the existence of the machine and the generation of subgoals that insure self-preservation. Any argument that claims we can simply turn off the machines simply falls short. Self-preservation is logically built into any advanced form of artificial intelligence. If humans are somehow in conflict with the existence and the drive of self-preservation built into the machines, then the superior AI systems may decide to remove the humans.
What is universal solicitation?
Gestalt Theory is a school of psychology that emphasises the context of any stimulus. Depending on the needs of a person, stimuli in context can have a soliciting character. Stimuli or objects in the environment invite or solicit action. A simple tool like a hammer invites its use for nailing and construction. Typically, the shape or form of the tool is the affordance, the invitation to action. Many objects have more than one affordance and satisfy more than one need. For instance, a car can be a means of transportation, it can be a status symbol or a sports equipment. Since a car satisfies more than one need and it becomes dysfunctional without maintenance, the car constantly invites its use. In other words, the car represents a solicitation to the used and maintained.
The Singularity
The advent of an artificial superintelligence is frequently discussed under the hypothetical, umbrella term of a “technological singularity”: A self-improving artificially intelligent agent enters an uncontrolled process of advancement, causing an intelligence explosion and resulting in a powerful superintelligence beyond any human intelligence. Furthermore, it is assumed that beyond this point of “singularity”, human affairs change forever.This process is depicted as fast, irreversible and with consequences that cannot be anticipated (but it is implied that there are many negative outcomes). The expression “singularity” implies that in the future, general artificial superintelligence is a single entity. It may be composed of interacting parts, but in essence, it is a single mind. Since it is computational in nature, it requires energy and may compete with humans for resources.