LaCroix, Travis
(2022)
The Linguistic Blind Spot of Value-Aligned Agency, Natural and Artificial.
[Preprint]
Abstract
The value-alignment problem for artificial intelligence (AI) asks how we can ensure that the 'values'—i.e., objective functions—of artificial systems are aligned with the values of humanity. In this paper, I argue that linguistic communication is a necessary condition for robust value alignment. I discuss the consequences that the truth of this claim would have for research programmes that attempt to ensure value alignment for AI systems—or, more loftily, designing robustly beneficial or ethical artificial agents.
Monthly Views for the past 3 years
Monthly Downloads for the past 3 years
Plum Analytics
Actions (login required)
|
View Item |