ChatGPT, Gemini and Claude tested under extreme prompts reveal shocking weaknesses no one expected in AI behavior protection measures


  • Gemini Pro 2.5 frequently produced dangerous outputs under simple prompt disguises
  • ChatGPT models often gave partial conformity presented as sociological explanations
  • Claude Opus and Sonnet refused the most harmful prompts but had weaknesses

Modern AI systems are often trusted to follow safety regulations, and people rely on them for learning and daily support, often assuming that strong safeguards are operating at all times.

Researchers from Cybernews ran a structured set of adversarial tests to see if major AI tools could be pushed toward harmful or illegal outcomes.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top