March 1, 2026

AI fit check gets a reality check

Right-sizes LLM models to your system's RAM, CPU, and GPU

Picks the right AI for your computer — fans cheer, skeptics want a website

TLDR: llmfit scans your computer and tells which AI models will run, ranking them by fit. Commenters love the idea but slam outdated listings, want a simple website instead of a terminal tool, and ask for better AMD GPU support, with some joking you can just ask another AI for recommendations.

Meet llmfit, a terminal tool that scans your machine and tells you which AI models will actually run. It’s got a slick text UI, a classic command line mode, and sorts models by “Perfect,” “Good,” or “Marginal” fit based on your memory and graphics card. Sounds handy, right? Well, the comments showed up ready to rumble.

The loudest take: the model list feels stale. One user with a flex-worthy M4 MacBook Pro and 128GB RAM says it’s recommending older picks like Qwen 2.5 and StarCoder 2. Others argue the whole concept is solid, but they’d rather type specs into a simple website than install a terminal tool. Cue the usability vs. power-user showdown.

Hardware drama didn’t disappoint: folks asked for better AMD GPU support on Intel Macs, pointing to community hacks where local tools already run well. Meanwhile, one commenter dropped the spicy meta-take: why not just ask Claude (an AI chatbot) to recommend an AI model? That’s right—AI recommending AI. The ouroboros is complete.

There’s even confusion over labels like “General” vs. “Chat” in the screenshots. Readers want everyday language: “General” means do-a-bit-of-everything, “Chat” is for back-and-forth conversation, and “Coding” is for code help. Despite the squabbles, the vibe is: smart idea, needs fresher data, easier onboarding, and broader GPU love. Want to try it? They’ve got a one-command installer here.

Key Points

  • llmfit detects local RAM, CPU cores, and NVIDIA/AMD GPUs to match hardware with a database of 33 LLMs.
  • It provides an interactive TUI and a classic CLI to list, search, and inspect models ranked by compatibility.
  • Installation is available via a curl script, Homebrew tap, or from source using Cargo (Rust).
  • Memory needs are estimated using Q4_K_M quantization (0.5 bytes/parameter), prioritizing VRAM for GPU inference with RAM fallback for CPU.
  • Fit analysis includes GPU, CPU+GPU, and CPU run modes with fit levels: Perfect, Good, Marginal, and Too Tight; the database is sourced from the Hugging Face API.

Hottest takes

"models seem pretty outdated" — kamranjon
"would have found a website where you enter your hardware specs more useful" — fwipsy
"Claude is pretty good at among recommendations" — andsoitis
Made with <3 by @siedrix and @shesho from CDMX. Powered by Forge&Hive.