There's an infinite way for me to kill bugs in the dirt in my back yard: Plant flowers, pave the driveway, play with dog in yard, etc.
NONE of them involve malice.
All of them are beyond the imagination of the bugs.
(This tweet is about AGI)
Here’s an
article about all the non-malicious ways AGI could treat us:
Argument 1: A Superintelligence Will be Super Emotionally Intelligent
Argument Summary:
“Such a massive superintelligence will naturally be compassionate, or humans will only build such a superintelligence, or - if such an intelligence if formed with cognitive enhancement - it will carry compassion with it.”
Rebuttals:
* To speak of what a superintelligence would “naturally” do is like a squirrel speaking of what a human will “naturally” do. The traits and behaviors and modes of valuing and acting that a superintelligence takes on will be wholly foreign and literally unintelligible to human beings. [...] While I’ve interviewed them and respect their ideas - I disagree quite strongly with Kornai’s notion that a “rational” machine would be moral in some way that is intelligible to humanity - and with Voss’s idea that improved intelligence means improved morality (in some human-comprehensible way).
* Not “natural” is compassion in the first place? Humans - with all our capacity for this touted virtue - have done and continue to do tremendous damage to ourselves and other species. [...] If nature shows us anything it is the predominance of the drive to survive (Spinoza’s conatus). Sometimes through fuzzy warm feelings, sometimes through avoiding predators, sometimes through injecting hosts with eggs so that babies may crawl out of the host’s head while it is alive, kicking and screaming. Nature has never rewarded virtue itself, but only actions that behoove the actor.
* There is literally no reason to believe that human values like “love” or “compassion” or “humor” will matter in any way to a superintelligence. Such an entity would inevitably have more complex drives and motives and means of metacognition than we possibly have. It may have entirely different levels of consciousness - or an ability to alter matter or time in ways that we can’t image - much like a mouse cannot imagine constructing an iPhone.
Argument 2: An AGI Child Will Revere its Parent
Summary:
“A superintelligence, created by man, would revere man, would revere its direct creators, or even the entire species that created it - so it would never want to harm humans.”
Rebuttals:
* Tell me how much Commodus revered Aurelius. How far the apple sometimes falls. Children are often not just indifferent to, but spiteful of, their parents. Machines with no brain chemicals to sense an emotional kinship seem even more likely to be indifferent to a “parent” (a term that is astronomically too anthropomorphic).
* A machine needn’t dislike humanity to (a) destroy it, as we destroy colonies of ants or acres of trees when we build a building or a road, or (b) wholly neglect it, leaving it to die off on its own. We have better things to than cater to the needs of ants, and superintelligence will have better things to do than cater to the needs of humanity. No harm and no malice required.
* A rebuttal to this rebuttal is that a machine could be built to value living things, and to value the happiness - and absence of suffering - of living things. Even if this were the case, there is no reason to believe that keeping physical humans alive and well would be the best way to achieve that goal. Mind uploading, or building utilitronium might be a much better ways to achieve this end. A mouse imagines travel to be running - humans invent hypersonic jets and land on the moon. We imagine “compassion” or “maximizing wellbeing” to be lots of happy little humans - but a superintelligence will have different ideas entirely, ideas that we can’t possibly imagine.
Argument 3: The Zoo Scenario
Summary:
“Even if a superintelligence comes to rule the world, it would want to keep humans alive, preserving our diversity, and observing us carefully as we humans observe other wildlife around us.”
Rebuttals:
* Anthropomorphic again. What leads us to believe that (a) we will be interesting or entertaining enough to keep around, or that (b) such a condition will be pleasant for humans, or (c) that superintelligence will - in any way - have curiosities like our own? These are preposterous suppositions.
* The zoo scenario is often based on the supposition that superintelligence would value biological diversity. First - why? If a superintelligence could create simulations of all known biological lift - and learn everything possible from these life forms, it may have no need for biology. Or, rather than keeping alive all the various forms of earth-life, it might simulate a billion earths with a billion different variations on various species, playing these simulations at a billion times the speed of time as we now experience it. AGI would “value” diversity in a way that we don’t possibly understand, if at all.
***
Strongest Rebuttal
The absolutely strongest rebuttal to all arguments of certainty that humans will be safe after the dawn of AGI is the following:
* None of us have any goddamned clue.
That’s the actual state of things. There are an unreasonably large number of possible minds and ramifications for those minds (Roman Yampolski’s 2016
speech on this topic is among the best on the topic). Suspecting that any of us know “what it would do” is preposterous. The camp of “rational machines will be naturally benevolent” or “machine will want to kill us all” are both ignorant of what a superintelligent machine would do. Any felt sense of certainty about the behavior or traits of superintelligence is about as reliable as a wasp’s understanding of the traits of human beings - we are wise crickets at best. A machine with vastly expanding intelligence is likely to go through tremendous change and evolution in its senses, its cognition, its abilities, and its ways of valuing and acting in the world (what we might call “ethics”, and what it will have a much more nuanced and robust understanding of). Some phases of these oscillations and expansions and alterations of intelligence and valuing are likely to devalue humans, or to neglect humans, and in those intervals, we may well be ignored or wiped out. As I mention in Moral Singularity, it’s somewhat ridiculous to suspect that an expanding alien mind, whose modes of valuing and deciding are also expanding, would indefinitely and continually arrive at valuing - simultaneously - humanity’s: happiness, autonomy, and wellbeing.
At the very least, it would be irrational to suspect that and ever-increasing superintelligence with ever-evolving and expanding modes of acting and valuing things would somehow always - for thousands or millions of years - place some particular and unique value in keeping hominids happy. I have written an entire essay - drafted in late 2012 - on this exact topic: Morality in a Transhuman Future - Repercussions for Humanity. I don’t consider the argument to be all that special, but I know of no strong argument to refute it. In my opinion, it remains the rebuttal of rebuttals to the “machines will always treat humans well” argument.
***
What We Do About It
I’ve argued that a global steering and transparency committee is probably the only way to prevent the advancement of strong AI and cognitive enhancement to lead to war. It seems to me that continued ethical thought on the matter - like that conducted by Yudkowsky and Hanson and many others - seems fruitful - even if AGI is many decades away (I still sometimes recommend people read the old
AI Foom debate between these two thinkers). While AI alignment is probably a critical problem for us to solve, we will ultimately have to grapple with the end-goal of humanity and the trajectory of intelligence itself - for Emerson didn’t suspect that the spire ends with man:
A subtle chain of countless rings
The next unto the farthest brings;
The eye reads omens where it goes,
And speaks all languages the rose;
And, striving to be man, the worm
Mounts through all the spires of form.
I don’t argue machine malice (“nastiness”) necessarily. If anything I think we will matter to it like ants matter to us, or like we matter to Spinoza’s indifferent god. I link (in the article) to this 2013 essay: On Morality in a Transhuman Future. There’s the TL;DR of the essay above, in order:
1. We could only predict or understand a superintelligence’s “morality” as well as a muskrat can understand and predict human morality (i.e. most of it will be vastly beyond our ability to understand).
2. An AI whose intelligence and understanding grows and swells (maybe fast with an AI Foom, maybe slow without one) would go through many different phases of development in how it values things (“morality”) and acts.
3. It is extremely unlikely that all of these phases of development you place real value on human beings, or that all phases would even consider us worthy of allocating attention or resources.
4. Some individual “phase” of a superintelligence’s moral development, or some combination of these phases, would almost inevitably lead to our destruction or our neglect and withering away. No “nastiness” or malice required.
I think that saying “an AGI would not be molded by biological survival and the State of Nature, so it would be less likely to express the viciousness of the State of Nature” is a reasonable statement. “Would” and “would not” seem far too certain, and while I can respect the position I can’t firmly grip that kind of certainty. The point about the state of nature, though, I think has credence. My intuition is that Omohundro is right about
AI Drives. However, he may well be wrong, and self-preservation may in fact not be My supposition is that, if strong AIs proliferate, the ones that “win” will share many of the same traits as animals that “win.” i.e. Spinoza’s conatus, or the core drive to survive and protect it’s own interest - by violence if need be. There might be some super-cooperation that would arrive, rather than super-competition, and there’s a big part of me that hopes for just that.