<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>How difficult is AI alignment? | Anthropic Research Salon</title>
        <link>https://tube.grossholtz.net/videos/watch/5e100e5e-71e0-46fa-a852-c1be08e13a4f</link>
        <description>At an Anthropic Research Salon event in San Francisco, four of our researchers—Alex Tamkin, Jan Leike, Amanda Askell and Josh Batson—discussed alignment science, interpretability, and the future of AI research. Further reading: Anthropic’s research: https://anthropic.com/research Claude’s character: https://www.anthropic.com/news/claude-character Evaluating feature steering: https://www.anthropic.com/research/evaluating-feature-steering 0:00 Introduction 0:30 An overview of alignment 4:48 Challenges of scaling 8:08 Role of interpretability 12:02 How models can help 14:31 Signs of whether alignment is easy or hard 18:28 Q&amp;A — Multi-agent deliberation 20:38 Q&amp;A — Model alignment epiphenomenon 23:43 Q&amp;A — What solving alignment could look like</description>
        <lastBuildDate>Mon, 06 Apr 2026 03:10:13 GMT</lastBuildDate>
        <docs>https://validator.w3.org/feed/docs/rss2.html</docs>
        <generator>PeerTube - https://tube.grossholtz.net</generator>
        <image>
            <title>How difficult is AI alignment? | Anthropic Research Salon</title>
            <url>https://tube.grossholtz.net/client/assets/images/icons/icon-512x512.png</url>
            <link>https://tube.grossholtz.net/videos/watch/5e100e5e-71e0-46fa-a852-c1be08e13a4f</link>
        </image>
        <copyright>All rights reserved, unless otherwise specified in the terms specified at https://tube.grossholtz.net/about and potential licenses granted by each content's rightholder.</copyright>
        <atom:link href="https://tube.grossholtz.net/feeds/video-comments.xml?videoId=5e100e5e-71e0-46fa-a852-c1be08e13a4f" rel="self" type="application/rss+xml"/>
    </channel>
</rss>