1. Signal‑to‑noise ratio – how useful the reviews actually are
Many users complain that AI reviewers churn out a lot of “nit‑picks” and miss the real bugs.
“The signal‑to‑noise ratio problem is unexpectedly difficult.” – dakshgupta
“I’ve been using it a bit lately and it’s pretty good, but it also exhibited strange behavior such as entirely overwriting PR descriptions with its own text.” – disillusionist
2. Human oversight – AI can’t replace a reviewer, it needs a human in the loop
Most participants agree that a human must decide whether to act on a comment, and that AI should be an assistant rather than a gatekeeper.
“You need a human in the loop to decide whether to act on a comment.” – tayo42
“The human is shifted right and can do other things rather than grinding through fiddly details.” – pnathan
3. Independence & integration – can the reviewer be truly separate from the generator?
Debate over whether a tool that uses the same LLM for both writing and reviewing is truly “independent” and how that affects trust.
“Independence is ridiculous – the underlying LLM models are too similar.” – sdenton4
“If the reviewer is part of the same system, it’s not an independent activity.” – rohansood15
4. Market saturation & business model – why pay for another AI review tool?
Participants note that many vendors are offering similar functionality, raising questions about differentiation and pricing.
“It’s a bubble – every vendor is shipping a code‑review agent.” – trjordan
“You could just ask Claude for a review and then distill this into PR comments.” – the__alchemist
These four themes capture the bulk of the discussion: how well the tools work, the need for human judgment, the technical independence of the reviewer, and the crowded, cost‑driven market.