Philosophical Investigations into AI Alignment: A Wittgensteinian Framework

被引:0
|
作者
José Antonio Pérez-Escobar
Deniz Sarikaya
机构
[1] École Normale Supérieure,Centre Cavaillès, UAR 3608 République Des Savoirs
[2] PSL University,Department of Philosophy
[3] Department of Logic,Centre for Logic and Philosophy of Science
[4] History and Philosophy of Science,Ethical Innovation Hub
[5] UNED,undefined
[6] University of Geneva,undefined
[7] Vrije Universiteit Brussel (VUB),undefined
[8] Universität zu Lübeck,undefined
关键词
Later Wittgenstein; Alignment Problem; AI safety; Meaning as use; Rule bending;
D O I
10.1007/s13347-024-00761-9
中图分类号
学科分类号
摘要
We argue that the later Wittgenstein’s philosophy of language and mathematics, substantially focused on rule-following, is relevant to understand and improve on the Artificial Intelligence (AI) alignment problem: his discussions on the categories that influence alignment between humans can inform about the categories that should be controlled to improve on the alignment problem when creating large data sets to be used by supervised and unsupervised learning algorithms, as well as when introducing hard coded guardrails for AI models. We cast these considerations in a model of human–human and human–machine alignment and sketch basic alignment strategies based on these categories and further reflections on rule-following like the notion of meaning as use. To sustain the validity of these considerations, we also show that successful techniques employed by AI safety researchers to better align new AI systems with our human goals are congruent with the stipulations that we derive from the later Wittgenstein’s philosophy. However, their application may benefit from the added specificities and stipulations of our framework: it extends on the current efforts and provides further, specific AI alignment techniques. Thus, we argue that the categories of the model and the core alignment strategies presented in this work can inform further AI alignment techniques.
引用
收藏
相关论文
共 50 条