We are discussing whether, in a society where there are several free
AI:s and superintelligences around, the superintelligences would need
to have human-friendly values in order for the humans to survive, or
whether pragmatic reasons alone would suffice to motivate the
superintelligences to preserve us.
Peter C. McCluskey wrote:
> >Hmm.... I suspect that superintelligences' ability to deal with a
> >complex law code will vastly excede our own, so adding a clause saying
> >that beings below a certain intelligence threshold (higher than
> >human-level) don't have property rights would seem like a manageble
>
> Most of us can handle a much more complex legal system than we do. There
> are plenty of advantages to having a simple legal system. A legal system
> which maximizes economic efficiency will probably make most intelligences
> richer than a less efficient one which maximizes their fraction of the
> current wealth.
Why do you think that a legal system that included the extra clause
(saying that human-level intelligences lack don't have rights) would
be less economically efficient than one which didn't? Consider:
(1) Once the humans are annihilated, nobody needs to think about the
law any longer.
(2) A line has to be drawn somewhere. Why is it inefficient to draw
it between humans and superintelligences, but not between, say,
slightly uplifted animals and humans?
(3) Maybe superintelligences can manage their capital more
efficiently than somewhat irrational humans.
> >complication. Especially if humans claimed the ownership of a
> >significant fraction of cosmos, or even just the solar system; then
> >it would seem worth the cost for the SIs to put in this extra clause
> >in their legal code.
>
> Yes, trying to claim ownership of things over which you have no control
> can undermine your ability to defend your property rights where they matter
> most.
And if humans are effectively powerless and don't have
direct "control" over anything?
> >Of course, the extra clause would only need to be temporarily
>
> I suspect you underestimate the costs of achieving agreement about a
> fundamental moral change.
Maybe. What does this cost consists in, and why do you think that it
will be big?
> >maintained, if the wasteful biological humans were promptly
> >exterminated and their belongings expropriated.
>
> One alternative that someone at the Foresight Senior Associates Gathering
> suggested: just upload all those humans in such a way that they aren't
> aware they've been uploaded. But that requires a bigger gap between humans
> and SI's than I forsee.
I think it might well be technologically possible, but it would still
imply coercion, since at least some of the original humans would
presumably have prefered to live in physical reality. Still, for us
humans at least, it seems like a much better outcome than
annihilation. If the singleton cares about human preferences (which
we will hopefully be able to make it do), then it would presumably
rather upload us than annihilate us. However, why stop there? Why
would the singleton not make an even greater effort to satisfy human
preferences by allowing those who so wish to continue to live in
physical reality (under some surveilance so they don't endanger the
singleton). It would cost somewhat more, but might still amount to a
very small fraction of the singleton's capital once space
colonization gets off the ground. Given that it is ethical in this
sense at all, chances are it's degree of "ethicality", as measured
by the fraction f of its capital that it devotes to humans, won't be
in the small interval given by
U<= f * (total capital) <=PR,
where U is the cost of keeping all humans uploaded, and
PR is the cost of also keeping those humans who so wish in physical
reality.
> As long as biological humans don't seriously interfere with a
> superintelligence's ability to achieve its goals, a small aesthetic
(or ethical, or sentimental--)
> bias built into that superintelligence may be enough to persuade it
> to coexist with lesser intelligences.
Yes.
>Reducing the costs of coexistence
> is probably easier than increasing the reliability of it's moral code.
Note that reducing the costs of coexistence is something *it*
does, whereas increasing the reliability of its code is something *we*
can do. So the former is not an alternative to the latter.
Also, see the above. Maybe the exact cost of coexistence doesn't
matter much. The question is then whether it will value coexistence
at all. (Though as long as it is restricted to this planet and its
near surroundings, then the cost of keeping humans in physical
reality might amount to a large fraction of its capital.)
_____________________________________________________
Nicholas Bostrom
Department of Philosophy, Logic and Scientific Method
London School of Economics
n.bostrom@lse.ac.uk
http://www.hedweb.com/nickb
Received on Tue Jun 2 02:36:10 1998
This archive was generated by hypermail 2.1.8 : Tue Mar 07 2006 - 14:45:30 PST