Value Alignment
Overview
We say we want AI to be value-aligned. But what exactly does this mean? An AI agent might act differently depending on whether it is aligned to its user's intentions, its user's preferences, its user's best interests, or overall moral rightness. This lecture gets students to consider these differences and reflect on whether different targets of alignment might be more or less appropriate for different contexts (for example, LLM chatbots). It also introduces them to the basics of some moral theories and considers the question of whether these theories offer us any help in creating value-aligned AI.
Contributors
- Ethics materials by Daniel Webber.
Ethics goals
- Understanding different conceptions of value alignment and the philosophical challenges they raise.