Skip to main content
eScholarship
Open Access Publications from the University of California

UC Santa Cruz

UC Santa Cruz Electronic Theses and Dissertations bannerUC Santa Cruz

Machine Learning and the Multiagent Alignment Problem

Creative Commons 'BY-SA' version 4.0 license
Abstract

In the context of artificial intelligence (AI) or machine learning (ML), we speak of the "alignment" of an AI system's behavior with human goals, values, and ethical principles. "The alignment problem" has proven challenging, and as the capabilities and applications of AI rapidly advance, the shortcomings of standard solutions are increasingly consequential. This dissertation focuses on an often overlooked but critically important complication to the alignment problem: Socially-consequential AI systems affect their environment (involving, for example, human populations) and are therefore subject to dynamical feedback driven by other agents. We address three central questions: (1) As intelligent agents adapt to each other, does a system aligned using current leading approaches remain aligned? (2) Can we anticipate and utilize adaptive agents' reactions to data-driven policy to achieve aligned objectives dynamically? (3) How can we guarantee alignment for AI systems that interact with complex, multiagent environments that are difficult to model or predict? We will address these questions using the theoretical framework and experimental tools of machine learning---integrating concepts from dynamical systems, evolutionary game theory, constrained optimization, and control theory. We hope to demonstrate that a dynamical systems approach to deployed AI is not only necessary but beneficial to the goal of alignment.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View