Philosophies (Dec 2020)

An AGI Modifying Its Utility Function in Violation of the Strong Orthogonality Thesis

  • James D. Miller,
  • Roman Yampolskiy,
  • Olle Häggström

DOI
https://doi.org/10.3390/philosophies5040040
Journal volume & issue
Vol. 5, no. 4
pp. 40 – 0

Abstract

Read online

An artificial general intelligence (AGI) might have an instrumental drive to modify its utility function to improve its ability to cooperate, bargain, promise, threaten, and resist and engage in blackmail. Such an AGI would necessarily have a utility function that was at least partially observable and that was influenced by how other agents chose to interact with it. This instrumental drive would conflict with the strong orthogonality thesis since the modifications would be influenced by the AGI’s intelligence. AGIs in highly competitive environments might converge to having nearly the same utility function, one optimized to favorably influencing other agents through game theory. Nothing in our analysis weakens arguments concerning the risks of AGI.

Keywords