Impossibility Results in AI: A Survey

被引:3
|
作者
Brcic, Mario [1 ]
Yampolskiy, Roman V. [2 ]
机构
[1] Univ Zagreb, Fac Elect Engn & Comp, Unska 3, Zagreb 10000, Croatia
[2] Univ Louisville, 132 Eastern Pkwy, Louisville, KY 40292 USA
关键词
Artificial intelligence; AI safety; limitations; impossibility theorems; NO-FREE-LUNCH; PHYSICAL LIMITS; INTELLIGENCE; EXISTENCE; THEOREMS;
D O I
10.1145/3603371
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
An impossibility theorem demonstrates that a particular problem or set of problems cannot be solved as described in the claim. Such theorems put limits on what is possible to do concerning artificial intelligence, especially the super-intelligent one. As such, these results serve as guidelines, reminders, and warnings to AI safety, AI policy, and governance researchers. These might enable solutions to some long-standing questions in the form of formalizing theories in the framework of constraint satisfaction without committing to one option. We strongly believe this to be the most prudent approach to long-term AI safety initiatives. In this article, we have categorized impossibility theorems applicable to AI into five mechanism-based categories: Deduction, indistinguishability, induction, tradeoffs, and intractability. We found that certain theorems are too specific or have implicit assumptions that limit application. Also, we added new results (theorems) such as the unfairness of explainability, the first explainability-related result in the induction category. The remaining results deal with misalignment between the clones and put a limit to the self-awareness of agents. We concluded that deductive impossibilities deny 100%-guarantees for security. In the end, we give some ideas that hold potential in explainability, controllability, value alignment, ethics, and group decision-making.
引用
收藏
页数:24
相关论文
共 50 条