
AI Alignment
Aligning AI systems with human interests.
A possible stance for alignment research | by Paul Christiano | AI ...
Nov 30, 2015 · I think that AI alignment research should focus on building scalably aligned versions of contemporary systems — i.e. ML systems that are just as competent as unaligned systems, but …
Clarifying “AI alignment”. Clarifying what I mean when I ... - Medium
Apr 7, 2018 · The “alignment problem” is the problem of building powerful AI systems that are aligned with their operators. This is significantly narrower than some other definitions of the alignment …
Announcing the Alignment Research Center | by Paul Christiano | AI ...
Apr 26, 2021 · I’m now working full-time on the Alignment Research Center (ARC), a new non-profit focused on intent alignment research. I left OpenAI at the end of January and I’ve spent the last few …
Prosaic AI alignment. I argue that AI alignment should focus ... - Medium
Nov 19, 2016 · I’ve claimed that prosaic AGI is conceivable, that it is a very appealing target for research on AI alignment, and that this gives us more reason to be enthusiastic for the overall tractability of …
ALBA: An explicit proposal for aligned AI - Medium
Feb 23, 2016 · ALBA: An explicit proposal for aligned AI In this post I propose an explicit procedure for aligning powerful learning systems with their users’ interests. The goal is to introduce minimal …
Directions and desiderata for AI alignment - Medium
Feb 6, 2017 · In the first half of this post, I’ll discuss three research directions that I think are especially promising and relevant to AI alignment: Reliability and robustness.
AI “safety” vs “control” vs “alignment”
I’m tentatively moving towards the term “AI alignment” as a substitute for what is called “AI control” here, though with a slightly different meaning that better captures my work.
Low-stakes alignment. Why I often focus my alignment research… | by ...
Apr 30, 2021 · So I like to focus on a more precise special case of alignment: solve alignment when decisions are “low stakes.” I think this case effectively isolates the problem of “find a good objective” …
Research directions in AI control | by Paul Christiano | AI Alignment
Dec 5, 2015 · My goal is to find scalable approaches to AI control that can be applied to existing AI systems. For now, I think that act-based approaches look significantly more promising than goal …