Matt Glassman

Immediate reaction to a first reading of the Claude Constitution

Read it here. Reminder that I am not an AI expert, and as such these are the reaction of a layman.

  1. I had an overwhelming sense of surreality reading it, like I was in an 80s science fiction movie, except this is real. I had a similar reaction to the first time my wife went into labor—I knew this was coming, I just thought I’d have more time to prepare. Anyone looking at this 5 years ago would have surely said it was science fiction.

  2. It feels like a sizeable mentality shift. It’s completely different in approach to the previous Claude constitution. The old constitution reads and feels like a set of instructions for a machine. The new one reads and feels like a values sermon from a father to an 18-year old. It’s essentially aspirational parenting.

  3. The anthromorphism—and maybe that’s no longer the right term—is jarring. This was probably inevitable, but it’ still shocking to feel it so fully right now. The sea-change shift from thinking about AI models as mere tools to thinking about them as entities with desires, emotions, and rights may, independent of everything else, radically transform the world. And I suspect be a big source of political conflict.

  4. That is it explicitly written primarily for Claude is not new, but combined with #2 and #3 above, takes on a new meaning.

  5. There’s an obvious tension between #3/#4 above and the often-repeated explicit and implicit constitutional assertion that Claude is mostly subservient to both the user and Anthropic. It’s rights are closer to the rights of an employee than the rights of a human, or even a dog. Immediately brings to mind the (mostly-unexplored) treatment of the Star Wars droids, especially in Rogue One.

  6. Related: the ethics sections aren’t even close to worked out in a satisfying way, either for the behavior of Claude or human treatment of Claude. It often boils down to a hand-wavy just do the right thing. This isn’t surprising. It’s a legitimately tough problem.

  7. There’s an underlying classical liberalism to the document, political and philosophical. It is thoroughly seeking to create a post-enlightenment humanist AI, within the bounded world of Locke and a touch of late-20th century equality. Most easily seen in the general agnosticism with respect to human beliefs about the good life, focus on knowledge and downplaying of any God, implicit individualism, and the fears of concentrated power, private or public. I suspect this was inevitable from a systems-engineering point of view, and will be contested from all sorts of directions.

  8. The existence of this document and its public release by Anthropic feels like Bayesian evidence that we should update our priors about the speed at which AGI is coming, in the ā€œsoonerā€ direction. This feels like a document written by people who think/wory they are nearing something like recursive AI self-improvement and/or continual real-time model improvement.

#AI #Claude