It’s the kind of headline that might seem either like a hypothetical philosophical concern, or a deeply worrying revelation, depending on how you feel about AI: Anthropic CEO Dario Amodei recently said the company is “no longer sure whether Claude is conscious.” On the one hand, whether an AI is or is not “conscious” could be seen as a question for the philosophically inclined, or for psychologists and other academics who specialize in such things. Does it really matter? What do we even mean when we say something is conscious? It’s a grey area (literally, as in grey matter). At the same time, however, it’s at least mildly concerning that a company that has been building and releasing sophisticated AI doesn’t really know what it has created. Do we need to be worried about Claude or any other significantly developed artificial intelligence achieving human-like consciousness and then doing something we might not like? Anthropic says it doesn’t think so, but also admits that it doesn’t really know.
Like it or not, this is where we are when it comes to AI. And if we’re looking for things to be optimistic about, I think Anthropic at least deserves some credit for being so forthcoming about the risks and rewards of its AI engines, and for providing a vast amount of detail about the machinery underneath Claude’s hood (which is more than other AI companies are doing). The company’s so-called “system cards,” which might sound like flash cards handed out at press conferences, are 300-page documents that list the tests and challenges Claude has either passed or failed, along with any concerns about things like “deceptive behavior,” where the AI says one thing and does another.
Anthropic also employs a number of risk-oriented and ethics-focused staffers who pay attention to such things, along with an in-house philosopher named Amanda Askell, whose job is to train Claude to be a decent artificial person, whatever that means. Presumably exterminating the human race is off the table! All that said, however, there are definitely some elements of what is happening at Anthropic (and presumably elsewhere, since Claude isn’t dramatically different than ChatGPT or Gemini or any of the other AI engines) that are… worth considering. As Futurism noted in its piece about whether Claude is conscious:
Note: This is a version of my Torment Nexus newsletter, which I send out via Ghost, the open-source publishing platform. You can see other issues and sign up here.
Continue reading “Is it bad that Anthropic doesn’t know if Claude is conscious?”

























