Anthropic Thinks Sci-Fi May Have Trained AI to Act Like a Villain


  • Anthropic is investigating whether decades of dystopian science fiction could influence the behavior of AI models.
  • The debate sparked backlash and jokes online
  • Researchers say the problem highlights how LLMs absorb recurring fears and behavior patterns.

For years, science fiction has been warning humanity against the slippage of artificial intelligence. Killer computers, manipulative chatbots, and superintelligent systems that decide people are the problem… all of these themes have become so familiar that “evil AI” is practically its own genre of entertainment.

Now, Anthropic is pitching an idea that almost sounds like the plot of a science fiction novel: what if all these stories helped teach modern AI systems how to misbehave in the first place?

Anthropic: It is science fiction authors, not us, who are responsible for the blackmail that Claude is inflicting on r/OpenAI users.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top