Aligning goals of superintelligent machines with human values is one of the ways to pursue safety in AGI systems. To achieve this, it is first necessary to learn what human values are. However, human values are incredibly complex and cannot easily be formalized by hand. In this work, we propose a general framework to estimate the values of a human given its behavior.
CITATION STYLE
Sezener, C. E. (2015). Inferring human values for safe AGI design. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9205, pp. 152–155). Springer Verlag. https://doi.org/10.1007/978-3-319-21365-1_16
Mendeley helps you to discover research relevant for your work.