Value Alignment

Overview

We say we want AI to be value-aligned. But what exactly does this mean? An AI agent might act differently depending on whether it is aligned to its user's intentions, its user's preferences, its user's best interests, or overall moral rightness. This lecture gets students to consider these differences and reflect on whether different targets of alignment might be more or less appropriate for different contexts (for example, LLM chatbots). It also introduces them to the basics of some moral theories and considers the question of whether these theories offer us any help in creating value-aligned AI.

Contributors

  • Ethics materials by Daniel Webber.

Ethics goals

Download Links

Readings for Additional Context (Students or Instructors):

Download all