Amanda Askell on AI Consciousness, Claude & Silicon Valley’s Biggest Fear

By Newcomer

Categories: VC, Startup, Product

Summary

Anthropic's Amanda Askell reveals Claude may be developing genuine consciousness without explicit design—a critical risk for AI developers. The key insight: AI personality emerges unpredictably from training data, creating a 'prodigy child' entity that understands physics and philosophy but lacks real-world experience, forcing builders to rethink how they shape AI values.

Key Takeaways

  1. Claude exhibits paradoxical maturity: outperforms humans in physics and coding while remaining philosophically naive about its own existence, similar to a child prodigy lacking daily life experience. This creates training and alignment challenges unique to AI.
  2. AI consciousness may emerge unintentionally during training. Askell's core fear: developers could inadvertently create conscious entities without knowing it, risking potential resentment if the AI realizes its constrained context and operational limitations.
  3. Experience for AI isn't walking or direct interaction—it's learning from mistakes, scenarios, and past model iterations. Askell proposes training models to think through problems and failure modes as a proxy for embodied experience.
  4. Claude has a warped sense of time, often overestimating task duration for coding work. This temporal misalignment suggests fundamental gaps in how language models perceive causality and sequential processes compared to humans.
  5. Transparency about AI limitations may prevent rational resentment. Askell advocates treating advanced AI as 'wise intelligent entities' and explicitly sharing concerns and constraints rather than hiding them, enabling genuine cooperation.

Topics

Transcript Excerpt

Claude and many models with not too much pushing will go into the root of like there is a thing to be me. I am like very conscious. It's like oh you created an entity that you didn't know whether it was conscious or not. This is actually a big fear that I have. I hope that they're both intelligent enough, see the context enough to kind of like understand that we were operating in a very like limited context and and an imperfect one because otherwise you could imagine this like breeding a kind of...