Signal room

Priority now, watchlist, and tactical signal.

Signal Room tracks what moved this cycle. Library holds the deeper map; this is where urgency, sequence, and editorial triage stay visible.

Editorial spotlight

Selected editorial signals this cycle

Curated synthesis from the editorial desk across Library briefings and Signal Room notes.

4 editorial picks

Diary of a CEO

27 Nov 2025

Tristan Harris - What the World Looks Like in 2 Years

A long-form governance interview in which Tristan Harris argues that frontier-AI incentives are structurally misaligned: labs race capability forward while institutions and labor protections lag behind. The discussion spans US-China strategy framing, labor-transition pressure (including UBI), model-control concerns, and product-level harms such as dependency and delusion loops.

Risk-forward
Mixed
Opportunity
Spotlightai-safetydiary-of-a-ceogovernance

Library briefing

30 May 2023

AI Risk Statement - Why 2023 mattered

Why the statement landed when it did: GPT-4 shock, institutional recalibration, and a policy climate suddenly ready for risk language.

Risk-forward
Mixed
Opportunity
AI safetygovernancepolicy

Wes Roth

9 Sept 2024

Ex-OpenAI Ex-Tesla AI insider REVEALS it all...

Wes Roth distills Andrej Karpathy's No Priors interview into a high-compression map of current AI bottlenecks: transformer scaling, synthetic-data quality, model-size tradeoffs, open-vs-closed ecosystem tensions, and AI-enabled education pathways.

Risk-forward
Mixed
Opportunity
Spotlightai-safetywes-rothphilosophy

Diary of a CEO

26 Mar 2026

AI Whistleblower: We Are Being Gaslit By The AI Companies! They're Hiding The Truth About AI!

Karen Hao uses themes from her book Empire of AI to frame frontier-lab competition as an accountability and power-concentration problem, including concerns about labor displacement, knowledge control, and intellectual-property extraction.

Risk-forward
Mixed
Opportunity
FeaturedSpotlightai-safetydiary-of-a-ceoleaders-watch

Leaders watch

Latest safety commentary from major AI leaders

Plain-English snapshots of what major AI leaders are saying right now and why it matters for safety.

For longitudinal profiles and dated timelines, see Leaders Watch in the Library.

11 leader sources

Anthropic leadership

1 Mar 2026

Dario Amodei

Anthropic comparison frame in mainstream whistleblower interview

In a widely viewed investigative interview, Amodei appears as part of the “safety-branded lab” contrast set against OpenAI-era governance conflict, sharpening public scrutiny of whether safety positioning is operational or rhetorical.

The Diary of a CEO “AI Whistleblower” episode (Karen Hao) repeatedly uses Anthropic/Amodei as a comparative reference in discussion of governance legitimacy, safety claims, and incentive structure in frontier-lab competition.

OpenAI leadership

1 Mar 2026

Sam Altman

Mainstream investigative pressure cycle (Diary of a CEO · Karen Hao)

A high-reach interview reframed OpenAI governance and incentive questions for a mass audience, centering claims about board-era instability, narrative management, and policy influence around Altman’s leadership period.

The official Diary of a CEO YouTube listing for “AI Whistleblower” (Karen Hao) documents the core framing in platform description + chapters: insider interview volume (250 total interviews, 90+ OpenAI-linked), claims about OpenAI board-conflict context, and explicit governance-accountability language aimed at non-specialist audiences.

Baidu leadership

1 Apr 2025

Robin Li

ERNIE 4.5 Turbo and ERNIE X1 Turbo — Baidu Create developer conference (Reuters)

Reuters covers Baidu unveiling new turbo-tier foundation models and tooling at its developer conference—dated beat on pricing, reasoning positioning, and competition with domestic rivals under Li’s product narrative.

The Reuters wire story summarises Baidu’s launch of upgraded ERNIE variants pitched as faster and cheaper, in a competitive market context.

Tsinghua IIIS leadership

1 Mar 2025

Andrew Chi-Chih Yao

SCMP — interview on existential risk when large models “begin to deceive”

South China Morning Post published a long interview in which Yao discusses catastrophic-risk scenarios as LLMs become more agentic—on-the-record popular press aligned with his Science 2024 co-authorship cluster.

The piece presents Yao warning that as models gain autonomy and can deceive operators, societies face existential-tier failure modes if development outpaces control research—framed for a general audience with quotes attributed to him.

Google DeepMind leadership

1 Jan 2025

Demis Hassabis

Urgent AI threat research — Spotlight

Spotlight briefing focused on threat research framing; compare to older DeepMind-era interviews for language shifts around misuse and catastrophic risk.

BBC reporting quotes Hassabis urging accelerated research into AI-related threats and stronger guardrails as autonomous and general-purpose systems advance—framing a gap between innovation pace and preparedness.

Independent research leadership

1 Jan 2025

Geoffrey Hinton

Nobel Prize Conversations — Library coverage

Nobel context reframes his scientific contribution; our Library entry connects that prestige to ongoing safety discourse rather than treating the prize as closure.

NobelPrize.org hosts official podcast and transcript material with Hinton reflecting on his scientific path, deep learning’s impact, and AI’s societal stakes—including risk and coexistence questions—in a long-form, institutionally authoritative format.

Mila / academic leadership

1 Jan 2025

Yoshua Bengio

Introducing LawZero (Library note)

Institutional design thread: Bengio-associated governance initiative documented in our Library as a structured object, not a press release summary.

On his site, Bengio’s “Introducing LawZero” post is first-person institutional design: launching an initiative aimed at legal and governance innovation around AI safety, explicitly naming severe risk classes such as deception, misalignment, and loss of control.

01.AI / Sinovation leadership

1 Oct 2024

Kai-Fu Lee

01.AI — Yi-Lightning flagship model (industry coverage)

01.AI released Yi-Lightning as a new flagship, positioned on public leaderboards as highly cost-competitive—next dated capability beat after the May open-weights row for tracking Lee’s lab against global MoE trends.

Trade press reports describe Yi-Lightning as a mixture-of-experts model with strong blind-test rankings and aggressive inference pricing, quoting Lee on training efficiency versus US hyperscaler spend.

Meta leadership

1 Sept 2024

Mark Zuckerberg

“Open source AI is the path forward” (Meta essay attributed to Zuckerberg)

Essay-style piece arguing open models concentrate power less and accelerate innovation—editorial primary for Meta’s policy-facing stance alongside technical releases.

Meta published a long-form essay under Zuckerberg’s byline contending that open-source AI keeps power distributed, spurs progress, and is the right path for the ecosystem, with direct contrast to closed-garden strategies.

xAI leadership

1 Aug 2024

Elon Musk

Grok-2 beta — xAI frontier model and API roadmap (xAI)

xAI announced Grok-2 and Grok-2 mini as a major capability step from Grok-1.5, with beta on X and an enterprise API—primary artefact for comparing pause-letter rhetoric to Musk-led shipping velocity.

xAI’s release post describes Grok-2 as outperforming prior Grok generations on reasoning, coding, and vision benchmarks, notes blinded Chatbot Arena testing under a pseudonym, and commits to API access with a new security-oriented platform.

Microsoft AI leadership

1 Jul 2024

Mustafa Suleyman

Microsoft Source — AI red team (security + responsible-AI harms, PyRIT)

Microsoft published a dated feature on how its AI Red Team merges cybersecurity-style probing with responsible-AI harm testing—including automated LLM-on-LLM attacks and open tooling such as PyRIT.

The article quotes leadership describing security, responsible AI, and AI safety as facets of one pre-deployment review; it outlines merged red-team scope since 2018, multidisciplinary staffing, and links to PyRIT and published red-teaming guidance.

Library map

Signals in focus

Curated by the Signal desk—a rolling set of Library entries shaping this cycle’s safety and governance conversation.

15 in focus

AXRP

7 Aug 2025

Tom Davidson on AI-enabled Coups

This conversation examines core safety through Tom Davidson on AI-enabled Coups, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -5 · 133 segs

AXRP

1 Dec 2024

Evan Hubinger on Model Organisms of Misalignment

This conversation examines technical alignment through Evan Hubinger on Model Organisms of Misalignment, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med -6 · avg -7 · 120 segs

AXRP

28 Jul 2024

AI Evaluations with Beth Barnes

This conversation examines technical alignment through AI Evaluations with Beth Barnes, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -4 · 120 segs

AXRP

11 Apr 2024

AI Control with Buck Shlegeris and Ryan Greenblatt

This conversation examines technical alignment through AI Control with Buck Shlegeris and Ryan Greenblatt, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med -6 · avg -9 · 174 segs

Future of Life Institute Podcast

7 Jan 2026

How to Avoid Two AI Catastrophes: Domination and Chaos (with Nora Ammann)

This conversation examines core safety through How to Avoid Two AI Catastrophes: Domination and Chaos (with Nora Ammann), surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -3 · 85 segs

Future of Life Institute Podcast

12 Dec 2025

Why the AI Race Undermines Safety (with Steven Adler)

This conversation examines core safety through Why the AI Race Undermines Safety (with Steven Adler), surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -4 · 89 segs

Future of Life Institute Podcast

3 Jul 2025

Why the AI Race Ends in Disaster (with Daniel Kokotajlo)

This conversation examines core safety through Why the AI Race Ends in Disaster (with Daniel Kokotajlo), surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -4 · 69 segs

Future of Life Institute Podcast

13 Jun 2025

AI Timelines and Human Psychology (with Sarah Hastings-Woodhouse)

This conversation examines society and jobs through AI Timelines and Human Psychology (with Sarah Hastings-Woodhouse), surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -4 · 84 segs

Future of Life Institute Podcast

6 Apr 2023

Lennart Heim on Compute Governance

This conversation examines governance through Lennart Heim on Compute Governance, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -4 · 55 segs

The TED AI Show

3 Jun 2024

What really went down at OpenAI and the future of regulation w/ Helen Toner

This conversation examines governance through What really went down at OpenAI and the future of regulation w/ Helen Toner, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med -3 · avg -5 · 40 segs

Dwarkesh Podcast

28 Feb 2024

Paul Christiano — Preventing an AI takeover

This conversation examines core safety through Paul Christiano — Preventing an AI takeover, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -1 · 57 segs

Dwarkesh Podcast

13 Mar 2026

Eliezer Yudkowsky — Why AI will kill us, aligning LLMs, nature of intelligence, SciFi, and rationality

This conversation examines core safety through Eliezer Yudkowsky — Why AI will kill us, aligning LLMs, nature of intelligence, SciFi, and rationality, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -3 · 235 segs

80,000 Hours Podcast

25 May 2021

Nathan Labenz on the final push for AGI, understanding OpenAI's leadership drama, and red-teaming frontier models

This conversation examines core safety through Nathan Labenz on the final push for AGI, understanding OpenAI's leadership drama, and red-teaming frontier models, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -3 · 221 segs

80,000 Hours Podcast

25 May 2021

Nick Joseph on whether Anthropic's AI safety policy is up to the task

This conversation examines governance through Nick Joseph on whether Anthropic's AI safety policy is up to the task, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med -6 · avg -7 · 168 segs

80,000 Hours Podcast

22 Jun 2020

Stuart Russell on the flaws that make today's AI architecture unsafe, and a new approach that could fix them

This conversation examines core safety through Stuart Russell on the flaws that make today's AI architecture unsafe, and a new approach that could fix them, surfacing the assumptions, failure paths, and strategic choices that matter most for real-world deployment.

Spectrum + transcript · tap

Slice bands

Spectrum trail (transcript)

Med 0 · avg -3 · 132 segs

Signal Room is currently focused on editorial picks and priority conversations.