Source: Persuasion
by Matt Lutz
“How can we build an AI whose motivations are aligned with human well-being? This is an area on which I have expertise, as a philosopher who specializes in the foundations of moral reasoning. Unfortunately, I’m pretty sure that AI alignment is impossible.” (04/16/26)
https://www.persuasion.community/p/ai-alignment-is-impossible