Topics

The recurring debates, disagreements, and positions across the archive.

Agents

For most of 2024 and 2025, "agents" was the word everybody used and nobody could ship. Daniela Amodei says it plainly: the term was thrown around so much it lost meaning. Then, in early 2026, somethin...

AGI

The leaders of the three frontier labs broadly agree that AGI is coming, but they disagree sharply on when and what it even means. Dario Amodei puts himself at 90% confidence that a "country of genius...

AI Safety

The most clarifying moment for AI safety in early 2026 was not a research paper but a political confrontation. In February, Anthropic's Dario Amodei refused the Pentagon's demand for unrestricted mili...

Alignment

All three frontier labs agree alignment is the central unsolved problem, but they disagree sharply on method. Dario Amodei left OpenAI specifically because he felt the institution lacked genuine convi...

Competition

The frontier AI labs talk about shared values, but the February 2026 Pentagon crisis exposed how quickly solidarity fractures under pressure. When the Department of War blacklisted Anthropic for refus...

Compute

The frontier lab leaders all agree the AI race is a compute race, but they diverge sharply on how much raw hardware matters versus what you do with it. Sam Altman frames the buildout in almost civiliz...

Geopolitics

On the US-China technology competition, the frontier lab leaders are broadly aligned. Anthropic's formal submission on the Commerce Department's diffusion rule (January 2025) argued that chip export c...

Governance

The most consequential governance fight so far played out over a single weekend in February 2026. Defense Secretary Pete Hegseth gave Anthropic a Friday deadline: let the Pentagon use Claude for any l...

Interpretability

The three frontier labs agree that understanding what happens inside AI models matters. They disagree, sometimes sharply, on how to get there. Anthropic treats interpretability as a foundational scien...

Near-Term Risk

The near-term risk conversation has moved from hypothetical to actuarial. In February 2026, Dario Amodei laid out the case most bluntly: entry-level white-collar jobs will be automated, software engin...

Open Source

OpenAI was literally founded on the premise that open-sourcing AI would benefit humanity. Ilya Sutskever has said as much directly: the original tactic was to release everything, run as a nonprofit, a...

Predictions

The leaders of the three frontier labs have converged on a strikingly similar timeline: AGI arriving somewhere between 2027 and 2035, with most bets clustering around the nearer end. Dario Amodei puts...

Regulation

These leaders all agree AI needs some form of oversight, but the specifics reveal deep fault lines. The most dramatic test case came in February 2026, when Dario Amodei refused to give the Pentagon un...

Scaling Laws

Dario Amodei traces his belief in scaling laws back to 2019 at OpenAI, when GPT-2 offered the first real evidence that feeding models more data and compute produced predictable gains in intelligence. ...

Superintelligence

Dario Amodei has become the most aggressive voice on timelines. In February 2026, he told Dwarkesh Patel he puts 90% odds on reaching what he calls "a country of geniuses in a data center" within ten ...