About 50 results
Open links in new tab
  1. AI Alignment

    Aligning AI systems with human interests.

  2. A possible stance for alignment research | by Paul Christiano | AI ...

    Nov 30, 2015 · I think that AI alignment research should focus on building scalably aligned versions of contemporary systems — i.e. ML systems that are just as competent as unaligned systems, but …

  3. Clarifying “AI alignment”. Clarifying what I mean when I ... - Medium

    Apr 7, 2018 · The “alignment problem” is the problem of building powerful AI systems that are aligned with their operators. This is significantly narrower than some other definitions of the alignment …

  4. Announcing the Alignment Research Center | by Paul Christiano | AI ...

    Apr 26, 2021 · I’m now working full-time on the Alignment Research Center (ARC), a new non-profit focused on intent alignment research. I left OpenAI at the end of January and I’ve spent the last few …

  5. Prosaic AI alignment. I argue that AI alignment should focus ... - Medium

    Nov 19, 2016 · I’ve claimed that prosaic AGI is conceivable, that it is a very appealing target for research on AI alignment, and that this gives us more reason to be enthusiastic for the overall tractability of …

  6. ALBA: An explicit proposal for aligned AI - Medium

    Feb 23, 2016 · ALBA: An explicit proposal for aligned AI In this post I propose an explicit procedure for aligning powerful learning systems with their users’ interests. The goal is to introduce minimal …

  7. Directions and desiderata for AI alignment - Medium

    Feb 6, 2017 · In the first half of this post, I’ll discuss three research directions that I think are especially promising and relevant to AI alignment: Reliability and robustness.

  8. AI “safety” vs “control” vs “alignment”

    I’m tentatively moving towards the term “AI alignment” as a substitute for what is called “AI control” here, though with a slightly different meaning that better captures my work.

  9. Low-stakes alignment. Why I often focus my alignment research… | by ...

    Apr 30, 2021 · So I like to focus on a more precise special case of alignment: solve alignment when decisions are “low stakes.” I think this case effectively isolates the problem of “find a good objective” …

  10. Research directions in AI control | by Paul Christiano | AI Alignment

    Dec 5, 2015 · My goal is to find scalable approaches to AI control that can be applied to existing AI systems. For now, I think that act-based approaches look significantly more promising than goal …