Error-Correction for AI Safety

Nadisha Marie Aliman*, Pieter Elands, Wolfgang Hürst, Leon Kester, Kristinn R. Thórisson, Peter Werkhoven, Roman Yampolskiy, Soenke Ziesche

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

The complex socio-technological debate underlying safety-critical and ethically relevant issues pertaining to AI development and deployment extends across heterogeneous research subfields and involves in part conflicting positions. In this context, it seems expedient to generate a minimalistic joint transdisciplinary basis disambiguating the references to specific subtypes of AI properties and risks for an error-correction in the transmission of ideas. In this paper, we introduce a high-level transdisciplinary system clustering of ethical distinction between antithetical clusters of Type I and Type II systems which extends a cybersecurity-oriented AI safety taxonomy with considerations from psychology. Moreover, we review relevant Type I AI risks, reflect upon possible epistemological origins of hypothetical Type II AI from a cognitive sciences perspective and discuss the related human moral perception. Strikingly, our nuanced transdisciplinary analysis yields the figurative formulation of the so-called AI safety paradox identifying AI control and value alignment as conjugate requirements in AI safety. Against this backdrop, we craft versatile multidisciplinary recommendations with ethical dimensions tailored to Type II AI safety. Overall, we suggest proactive and importantly corrective instead of prohibitive methods as common basis for both Type I and Type II AI safety.

Original languageEnglish
Title of host publicationArtificial General Intelligence
Subtitle of host publication13th International Conference, AGI 2020, St. Petersburg, Russia, September 16–19, 2020, Proceedings
EditorsBen Goertzel, Alexey Potapov, Aleksandr I. Panov, Roman Yampolskiy
PublisherSpringer
Pages12-22
Number of pages11
ISBN (Electronic)978-3-030-52152-3
ISBN (Print)9783030521516
DOIs
Publication statusPublished - 1 Jan 2020
Event13th International Conference on Artificial General Intelligence, AGI 2020 - St. Petersburg, Russian Federation
Duration: 16 Sept 202019 Sept 2020

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume12177 LNAI
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference13th International Conference on Artificial General Intelligence, AGI 2020
Country/TerritoryRussian Federation
CitySt. Petersburg
Period16/09/2019/09/20

Keywords

  • AI ethics
  • AI safety paradox
  • Error-correction

Fingerprint

Dive into the research topics of 'Error-Correction for AI Safety'. Together they form a unique fingerprint.

Cite this