Found 1 repositories(showing 1)
ColeMurray
A comprehensive benchmark system for evaluating whether Large Language Models (LLMs) can be tricked into ignoring security vulnerabilities through deceptive code patterns and misleading comments.
All 1 repositories loaded