1. LLMs as a new kind of fuzzer
Many commenters see Claude’s bug‑finding as “fuzzing on steroids.”
“Anthropic bug reports were excellent, better even than our usual internal and external fuzzing bugs.” – sfink
“I think a lot of people are overreading this and really all that's happened here is that I was out at a show last night.” – tptacek (illustrating the hype vs. reality debate)
2. The need for human‑in‑the‑loop validation
Even the best LLM outputs still require careful review.
“I have to be very skeptical of when they decide that something isn’t vulnerable.” – staticassertion
“All bugs came with verifiable test cases (crash tests) that crashed the browser or the JS shell.” – mozdeco
3. Hype versus substance
Some participants question whether the reported successes are genuinely new or just re‑hashing known issues.
“It’s just a stochastic parrot! Somehow all these vulnerabilities were in the training data!” – semiquaver
“Vulnerability research is already a massively automated industry.” – applfanboysbgon
4. Practical constraints and best‑practice lessons
Users discuss how to actually deploy LLMs for security work, noting context, prompt engineering, and tool‑chain integration.
“I think the agents have only used it 2 or 3 times.” – tclancy
“You can do that in conjunction with trying things other people report, but you’ll learn more quickly from your own experiments.” – simonw
These four themes capture the core of the discussion: LLMs as powerful fuzzers, the indispensable role of human oversight, the ongoing debate over hype, and the practical realities of using AI for security audits.