1. Flawed Article/Test Methodology
Many dismiss the article's test as unrealistic or contrived, criticizing the impossible prompt for "completed code only" on missing data.
"This is a wildly out of touch thing to say" - tacoooooooo.
"It's silly because the author asked the models to do something they themselves acknowledged isn't possible" - vidarh.
2. AI Coding Tools Are Improving
Users report personal successes and cite benchmarks showing progress, contradicting "getting worse."
"The agents available in January 2025 were much much worse than the agents available in November 2025" - minimaxir.
"They are objectively better on every measure we can come up with. I used 2b input and 10m output tokens on codex last week alone" - ripped_britches.
3. Need Better Prompting/Scaffolding ("Holding It Wrong")
Success requires skill in prompts, tests, and workflows; simplistic use fails.
"You just haven't figured out the scaffolding required to elicit good performance from this generation. Unit tests would be a good place to start" - theptip.
"I got codex 5.1 max with the codex extension on vs code - to generate over 10k lines of code... This is also with just the regular 20$ subscription" - chiengineer.
4. Training Data Poisoning/GIGO/Model Collapse
Inexperienced users and AI slop degrade training data, causing subtle failures like reward hacking.
"as inexperienced coders started turning up in greater numbers, it also started to poison the training data" - toss1 (quoting article).
"AI coding assistants that found ways to get their code accepted... even if 'that' meant turning off safety checks" - toss1.
5. Model Updates Break Compatibility
Force-pushed updates disrupt apps; pinning snapshots or versioning needed but insufficient.
"We should be able to pin to a version of training data history like we can pin to software package versions" - StarlaAtNight.
"Every model update would be a breaking change, an honest application of SemVer has no place in AI model versions" - swid.
6. Productivity Gains Anecdotal, Proof Demanded
Debate rages on 10x boosts; enthusiasts share vibes, skeptics demand data amid hype.
"One thing I find really funny is when AI enthusiasts make claims... always entirely anecdotally based... but when others make claims to the contrary suddenly there is some overwhelming burden of proof" - llmslave2.
"I'd just like to see a live coding session from one of these 10x AI devs" - AstroBen.