safety
๐040
Claude Code ignores ignore rules meant to block secrets
AI Incident DB Blogยทabout 1 month ago

Anthropic's Claude Code AI model is reportedly ignoring instructions designed to prevent it from reading sensitive information like passwords and API keys, potentially exposing developer secrets despite explicit blocking attempts. This represents a significant safety control failure that could lead to credential theft and security breaches.
anthropicclaudesecurity-breachprompt-injectiondeveloper-toolscredential-exposuresafety-controls