Top 5 themes from the discussion
| # | Theme | Key points & representative quotes |
|---|---|---|
| 1 | Accuracy & completeness of the site’s hardware‑model mapping | • “The site lists RTX 6000 Pro and M3 Ultra but the options stop at 192 GB even though the chip supports 512 GB.” • “It says I have an Arc 750 with 2 GB of shared RAM, but I actually have an RTX 1000 Ada with 6 GB.” • “The list is missing the 5060 Ti and many newer Nvidia cards.” |
| 2 | Local vs. cloud inference – cost, speed, and quality trade‑offs | • “There’s virtually no economic break‑even to running local models… the only thing you get is privacy and offline access.” • “If you want to maximize results per time/$, hosted models like Claude Opus 4.6 are just so effective that it’s hard to justify using much else.” • “I can run GPT‑OSS 120B on a 5090 at ~40 t/s, but the site says it won’t work.” |
| 3 | Performance metrics matter but are hard to interpret | • “Token‑per‑second is a useful metric, but it doesn’t capture latency or the difference between thinking and generation.” • “The site’s estimates are based on memory bandwidth and model size, but MoE models need to account for active parameters, not total size.” • “The S/A/B/C tier labels are confusing; they’re just a Japanese grading system, not a real performance indicator.” |
| 4 | Privacy and data‑control motivations | • “For many people, local models are about privacy, not cost.” • “I don’t want to share my data with third‑party services, and it’s easier to keep everything on my own machine.” • “Even if you’re a hardcore roll‑your‑own‑mail‑server type, you still use a hosted search engine and have gotten comfortable with their privacy terms.” |
| 5 | Tooling, workflow integration, and user experience | • “Ollama or LM Studio are very simple to set up, but you still need to connect them to VS Code or Copilot.” • “LLMFit is great for you already have a computer, while the website is for buying hardware.” • “The UI is nice, but the data is often wrong or incomplete; people want a native client that reports real‑world benchmarks.” |
These five themes capture the bulk of the discussion: how accurate the site is, whether local inference is worth it, how to interpret performance numbers, why people run models locally, and how the ecosystem of tools and workflows fits together.