TL;DR — I gave Claude 4.5 a Kali box and an intentionally vulnerable app. In 15 minutes it produced a report with 21 real vulnerabilities (SQLi, exposed .git, misconfigured cookies), but it missed obvious XSS and some business logic issues. In the post I walk through the setup, what worked, what didn’t, and where AI actually belongs in a security workflow — useful for early dev checks and teaching, but not a replacement for manual pen testing. Read the full write-up: https://lnkd.in/gY7MzupX #AIsecurity #PenTesting #Infosec #Claude45
I agree — AI is essential for fast, scalable detection and routine containment, but the safest and most effective security posture outside a lab is a hybrid model: automate what’s repeatable and low-risk, and keep humans in the loop for high-impact judgment, oversight, and novel threats.
Would be interesting to do the same exact test with codex to see how it compares. But I've seen threat actors using HexStrike MCP to develop exploit for n-days so things can definitely get wild.
Aaron Ott Neat write-up, thanks for sharing! For your custom vulnerable app, I'm curious about the languages/web framework used, total lines of code, general complexity, etc. Also, I'm curious about the true positive/false positive rates on the app, and how consistently it finds the same bugs on subsequent scans. Regardless, cool work! :)
Now you can run it against a modern app, with captcha on login form and whole app is under Cloudflare WAF and has vulnerabilites inside authenticated areas. Codex or Claude will do nothing, they are only good on DVWA..
I also gave Claude 4.5 a vulnerable box. It told me it could not help me hack systems and did not answer my queries.
Gonna be honest, I first went "Ughh, another AI post" then read your blog and went "Hmmm let me give it a try" I like it and see use of this. I specially like the fact that Claude is asking for my permission before running commands. I'm allow listing most commands for this specific directory but it's feeling quite safe compared to most AI agents that decide to redo whole project's code on a whim. I can at least see this being useful for standard checks during pentest before the human gets creative.
Interesting results, I've been playing with similar ideas recently and there's some really cool capability potential to unlock here. CyberAgent and CAI are some examples on GitHub of how others have been implementing this idea.
The "intentionally vulnerable web app" was 100% custom made or DVWA/similar?
Is there a possibility that Claude referenced existing solutions or write-ups from the internet on DVWA?
Cybersecurity and AI in Robotics
2wHey Aaron, good one! I’d be interested in what’s your take on our open source Cybersecurity AI (CAI) https://github.com/aliasrobotics/cai and the supporting Cybersecurity LLM “alias1” https://aliasrobotics.com/alias1.php, which is an alternative to Anthropics’ Claude. You will obtain same results for a fraction of the cost and without so many refusals. Let me know if you’d like to try it out, happy to facilitate it