Project ideas from Hacker News discussions.

Using LLMs at Oxide

📝 Discussion Summary (Click to expand)

Here are the three most prevalent themes from the Hacker News discussion:

1. Concerns Over Junior Developer Development and Skill Atrophy

There is significant discussion about how reliance on LLMs might inhibit the fundamental learning and skill acquisition necessary for junior developers, drawing parallels to previous tools like auto-complete or Dreamweaver. The worry is that by abstracting away necessary struggles, developers won't build deep comprehension.

  • Supporting Quote: "For a junior in the learning phase that can be useful time spent. Then again, I agree that at times certain menial code tasks are not worth doing and llms are helpful. It's a bit like a kid not spending time memorizing their time tables since they can use a calculator. They are less likely to become a great mathematician." ("aatd86")
  • Supporting Quote: "I always typed it out again. Slower, but I swear it built the comprehension I have today." ("girvo")

2. The Imperative of Human Oversight and Ownership of LLM Output

A consensus emerges that while LLMs generate code or text, the human user must take ultimate responsibility for verification, review, and understanding—especially for code hitting production. The LLM is seen as a tireless assistant, but the engineer must remain the final arbiter of quality and correctness.

  • Supporting Quote: "The ultimate conclusion seems to be one that leaves it to personal responsibility - the user of the LLM is responsible for ensuring the LLM has done its job correctly." ("devmor")
  • Supporting Quote: "THOU SHALT OWN THE CODE THAT THOU DOST RENDER. All other values should flow from that, regardless of whether the code itself is written by you or AI or by your dog." ("keeda")

3. The Trade-off Between Craftsmanship/Quality and Speed/Productivity

Commenters frequently debated whether LLMs facilitate genuine engineering craft or merely produce "good enough" code quickly, leading to a tension between experienced developers who value deep understanding ("craftsmen") and those prioritizing rapid output ("practical types").

  • Supporting Quote: "I think there are two kinds of developers: craftsmen and practical “does it get the job done” types. I’m the former. The latter seem to be what makes the world go round." ("christophilus")
  • Supporting Quote: "The craftsman answer might be: use LLMs for the systematic/tedious parts (code generation, pattern matching, boilerplate) while keeping human judgment for the parts that require actual thinking. Let the tool handle what it's good at, you handle what requires actual thinking." ("govping")

🚀 Project Ideas

LLM Code Quality & Detailing Feedback Tool (CodeCraft Auditor)

Summary

  • A tool that audits LLM-generated code snippets, specifically targeting verbosity, style adherence, and over-reliance on implementation details (like excessive mocking in tests), providing actionable feedback before the engineer submits it for peer review.
  • Core value proposition: Reduce the review burden caused by LLM "slop" and ensure code ownership commitment by providing immediate, precise feedback on generated artifacts.

Details

Key Value
Target Audience Senior/Staff Engineers supervising junior developers or those using LLMs for large feature generation (e.g., the user who lamented reviewing "machine slop").
Core Feature Analyzes diffs containing LLM code against project conventions (style, complexity, test boilerplate) and generates specific refactoring suggestions (e.g., "Collapse these 5 mocks into a single behavior test," or "Reduce documentation verbosity by 50% based on established repo tone.").
Tech Stack Python/Rust backend, leveraging static analysis tools (e.g., AST parsers, linters) combined with fine-tuned open-source LLMs (or API calls to advanced models) specifically instructed on the project's established code style and common LLM anti-patterns (excessive use of em-dashes, test structure, etc.).
Difficulty Medium (Requires deep understanding of code transformation and effective prompt engineering to guide the auditing LLM accurately.)
Monetization Hobby

Notes

  • This directly addresses frustrations like: "I'm not entirely convinced that the autocomplete part of it is the best one," and the issue of LLMs producing code that is "5x and 20x as long as it needs to be."
  • It operationalizes the notion from the discussion: "Wherever LLM-generated code is used, it becomes the responsibility of the engineer." By providing feedback before commit/PR, it forces the engineer to complete their self-review step (#6 in john01dav's workflow).

Junior Developer LLM Alignment & Learning Monitor (Apprentice Compass)

Summary

  • A structured service/plugin that tracks the type of LLM assistance junior developers are using (autocomplete vs. generation vs. planning) and provides personalized coaching/gaps analysis compared to senior developer workflows.
  • Core value proposition: Solves the training dilemma: how to use LLMs effectively without causing atrophy ("muscle atrophy of the mind") in junior developers, ensuring comprehension over sheer output.

Details

Key Value
Target Audience Engineering Managers and Senior Developers responsible for training/mentoring junior staff.
Core Feature Integrates with IDEs/Source Control to log LLM interactions. It analyzes if the work done relied too heavily on LLM-generated solutions for foundational concepts (like creating a parser when one should build it manually for learning) and flags potential skill gaps to the mentor.
Tech Stack IDE extension (VS Code/JetBrains), telemetry collection, anonymized aggregated reporting dashboard (e.g., using a framework like FastAPI/React).
Difficulty High (Requires delicate instrumentation of developer workflows without creating excessive overhead or violating privacy/trust.)
Monetization Hobby

Notes

  • Addresses the key concern: "how do we train junior devs in that environment? Force them to build slowly, without llms?" This tool provides data-driven insight into where the assistance is happening, allowing managers to guide juniors toward "knowing what to look for" vs. rote generation.
  • It supports managers who value "comprehension any day" while acknowledging tools are necessary in the modern environment.

Context-Aware Codebase Idiom Extractor (Idiom Weaver)

Summary

  • A self-contained service that analyzes an entire, mature codebase (or a subset thereof) and generates an explicit, evolving "Project Style Guide" focused on deep structural idioms, semantic connections, and architectural preferences beyond surface-level syntax.
  • Core value proposition: Captures the tacit, high-level knowledge that LLMs struggle with ("deeper idioms present in a codebase") and formalizes it for both human onboarding and LLM context injection.

Details

Key Value
Target Audience Architects, core platform teams, and companies maintaining long-lived, complex software (like the Oxide team mentioned).
Core Feature Ingests the entire repo, runs advanced graph analysis to map semantic relationships, and outputs an LLM-ready specification (like an extended README or a structured JSON file) detailing non-obvious design constraints (e.g., "Data layer objects should never directly call serialization utilities; use the CodecManager interface").
Tech Stack Static analysis engines (e.g., custom parsers for target languages), graph databases (Neo4j) for relationship mapping, and an LLM specialized in synthesizing technical documentation from structured analysis output.
Difficulty High (Requires significant effort to establish the deep semantic mapping across a large codebase.)
Monetization Hobby

Notes

  • Directly addresses this pain point: "I have yet to get an LLM to pick up on the deeper idioms present in a codebase that help constrain the overall solution towards these more particular patterns."
  • This tool provides the necessary high-quality, deep context needed to move LLM usage beyond simple function generation into true feature writing while maintaining structural integrity.