From: Superintelligence
Indirect normativity offers a possibility to offload the question of which values to feed an agent with to the superintelligence itself while still anchoring it in deep human values.
for now a appropriate moral norm could be enough eg. "the common good principle": superintelligence should be developed only for the benefit of all of humanity and in the service of widely shared ethical ideals.; could include "windfall clause" for wealth distribution
Indirect normativity (256-259) Instead of directly determining ourselves which values an ai is to promote, humans specify a criterion or method that the ai can follow using its own intellectual resources to discover the concrete content of a mercy implicitly defined normative standard. "Since the superintelligence is better at cognitive work than humans it might see past the errors and confusion that cloud our thinking."
On human values and their change over time: "When we look back, we see glaring deficiencies not just in the behaviour but in the moral beliefs of all previous ages. Though we have perhaps since gleaned some moral insight, we could hardly claim to be now basking in the high noon of perfect moral enlightenment. Very likely we are still labouring under one or more grave moral misconceptions.
Seed ai might be given the final goal of continuously acting according to its best prediction of what the implicitly defined standard would have it do.
Coherent extrapolated volition from Eliezer Yudkowsky; seed ai is given final goal of carrying out humanities cev. cev: our wish if we knew more, thought faster, were more the people we wished we were, had grown farther together, where the extrapolation converges rather than diverges, where our wishes cohere rather than interfere, extrapolated as we wish that extrapolated, interpreted as wish that interpreted. –> compare ethics observer theories; trying to normative things like "good" and "right" –> only "take action" if wishes of all steak holders cohere; does not have to produce common ground, only find it where it exists -> further explanation on p. 260 Yudkowsky seven arguments for cev:
many open parameters remain eg. who's volitions are to be included; how to deal with "marginal persons" (p.265)
Morality models
"Do what i mean" –> due to possible misunderstandings in our own assertion unintended consequences can not be prohibited –> a more granular approach by clearing up the term "do what i mean" into revealed preferences in various hypothetical situations under the "what if..." model which ultimately leads back to the indirect normative cev approach
Other:
almost all the things we humans value (love, happiness, even survival) are important to us because we have particular evolutionary history – a history we share with higher animals, but not with computer programs, such as artificial intelligences.
http://theconversation.com/artificial-intelligence-can-we-keep-it-in-the-box-8541
––––––––
"So what we probably want is not a direct specification of values, but rather some algorithm for what's called indirect normativity. Rather than programming the AI with some list of ultimate values we're currently fond of, we instead program the AI with some process for learning what ultimate values it should have, before it starts reshaping the world according to those values."
https://io9.gizmodo.com/can-we-build-an-artificial-superintelligence-that-wont-1501869007
–––––––––
–––––––
3 principles for creating safer AI | Stuart Russell
–––––––
Sample Values from 23 ASILOMAR AI PRINCIPLES:
12) Personal Privacy: People should have the right to access, manage and control the data they generate, given AI systems’ power to analyze and utilize that data.
13) Liberty and Privacy: The application of AI to personal data must not unreasonably curtail people’s real or perceived liberty.
14) Shared Benefit: AI technologies should benefit and empower as many people as possible.
15) Shared Prosperity: The economic prosperity created by AI should be shared broadly, to benefit all of humanity.
16) Human Control: Humans should choose how and whether to delegate decisions to AI systems, to accomplish human-chosen objectives.
17) Non-subversion: The power conferred by control of highly advanced AI systems should respect and improve, rather than subvert, the social and civic processes on which the health of society depends.
––––––
http://intelligence.org/files/IE-ME.pdf
Value extrapolation theories have some advantages when seeking a machine ethics suitable for a machine superoptimizer:
––––––
https://nickbostrom.com/ethics/artificial-intelligence.pdf
To build an AI that acts safely while acting in many domains, with many consequences, including problems the engineers never explicitly envisioned, one must specify good behavior in such terms as “X such that the consequence of X is not harmful to humans”
Thus the discipline of AI ethics, especially as applied to AGI, is likely to differ fundamentally from the ethical discipline of noncognitive technologies, in that: