Asymptotically unambitious artificial general intelligence

10Citations
Citations of this article
30Readers
Mendeley users who have this article in their library.

Abstract

General intelligence, the ability to solve arbitrary solvable problems, is supposed by many to be artificially constructible. Narrow intelligence, the ability to solve a given particularly difficult problem, has seen impressive recent development. Notable examples include self-driving cars, Go engines, image classifiers, and translators. Artificial General Intelligence (AGI) presents dangers that narrow intelligence does not: if something smarter than us across every domain were indifferent to our concerns, it would be an existential threat to humanity, just as we threaten many species despite no ill will. Even the theory of how to maintain the alignment of an AGI’s goals with our own has proven highly elusive. We present the first algorithm we are aware of for asymptotically unambitious AGI, where “unambitiousness” includes not seeking arbitrary power. Thus, we identify an exception to the Instrumental Convergence Thesis, which is roughly that by default, an AGI would seek power, including over us.

Cite

CITATION STYLE

APA

Cohen, M. K., Vellambi, B., & Hutter, M. (2020). Asymptotically unambitious artificial general intelligence. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 2467–2476). AAAI press. https://doi.org/10.1609/aaai.v34i03.5628

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free