The Intergalactic Guide to LLM Parameter Sizes
Which AI Brain is Right for Your Mission? This is the absurdly over-complicated field guide that nobody asked for but everyone desperately needs
Recently I found myself on ollama.com looking at the options to run DeepSeek-R1 locally, when I encountered the digital equivalent of decision paralysis. The drop-down menu presented me with SEVEN different sizes of essentially the same model, ranging from a modest 1.5B all the way up to a completely ridiculous 671B parameters. To be sure, this isn't just a DeepSeek problem – you'll find the same array of options with Google's new Gemma3, Microsoft's Phi, Meta's big zoo of LLamas, and virtually every other model family on the market. When it comes to paramater sizes, selecting the right one felt less like choosing an AI model and more like being asked to specify the exact molecular weight of my next meal.
The storage requirements alone told a story of madness – everything from a reasonable 1GB to a staggering 404GB. That's not a download, it's a commitment... a relationship. Who has that kind of disk space to casually dedicate to a single model? And more importantly, who needs that many options of the same fundamental architecture?
After looking at my screen for longer than I'd like to admit, I realized what the world desperately needs: a straightforward, no-nonsense guide to this parameter size circus. In fact, I shouldn't even be the one writing it, neither do I see myself as that absolute authority for this topic, but we need that so let's just give it a go and I'll appreciate comments. So here's my attempt to decode this numerical madness without requiring a degree in computer science or making you want to abandon technology altogether. Just do not take it too seriously.
Tiny Models (1B-3B parameters): The Pocket Calculators
Size on disk: ~1-2GB
Hardware needs: Your grandma's laptop could run this
Power consumption: A hamster on a wheel could generate enough electricity
Actual usefulness: More than you'd expect, less than the marketing suggests
These models are like that pocket knife with only three tools—surprisingly handy despite obvious limitations. They're good for:
Just don't ask them to understand jokes, follow complex instructions, or remember what they said three messages ago. They've got the memory of a goldfish and the creativity of a tax form.
Small Models (4B-8B parameters): The Swiss Army Knives
Size on disk: ~3-5GB
Hardware needs: Any laptop made after the invention of TikTok
Power consumption: Roughly equivalent to a desk fan
Actual usefulness: The sweet spot for most people who don't work at an AI lab
The 7B size has become the unofficial "we made it just big enough to be useful" standard. These models can:
This is the realm where most people should start. It's like buying the mid-tier iPhone instead of selling your kidney for the Pro Max.
Medium Models (10B-20B parameters): The Desktop Computers
Size on disk: ~8-15GB
Hardware needs: Something with a GPU that doesn't catch fire
Power consumption: Your electricity bill will notice, but not scream
Actual usefulness: When you need to impress people but can't afford the big guns
These models occupy the awkward teenage phase of AI—not small enough to run easily everywhere, not large enough to blow minds. They offer:
The performance jump from 7B to 13B is often more noticeable than from 13B to 30B, making this a surprisingly practical choice if you can handle the hardware requirements.
Large Models (30B-70B parameters): The Workstations
Size on disk: ~20-40GB
Hardware needs: Gaming PC or better, preferably with an M3/M4 chip from Apple, or an NVIDIA card that cost more than your first car
Power consumption: Comparable to a small heater
Actual usefulness: The point where people start saying "wow" instead of "hmm"
Now we're talking serious horsepower—these models actually deliver on many of the promises made in AI marketing materials:
This is where the basic models of most commercial services like ChatGPT and Claude operate. There's a reason these aren't running on your phone.
Enormous Models (100B-200B parameters): The Server Racks
Size on disk: ~60-150GB
Hardware needs: Multiple high-end GPUs in a dedicated setup
Power consumption: Hope you've got solar panels
Actual usefulness: Overkill for 95% of use cases, but that remaining 5% is impressive
These behemoths are the AI equivalent of bringing a tank to a bicycle race—complete overkill for most situations, but undeniably powerful:
Unless you're doing cutting-edge research or running a commercial service, you probably don't need this. It's like buying a commercial espresso machine for your home when you drink coffee twice a month. Run those on a GPU-as-a-service infrastructure like replicate.com, together.ai or use the proprietary models via API.
Apocalypse-Inducing Models (500B+ parameters): The Supercomputers
Size on disk: 300GB+ (hope you've got fiber internet)
Hardware needs: Data center infrastructure, cooling systems, possibly a nuclear power plant
Power consumption: Comparable to a small town (I'm kidding)
Actual usefulness: Bragging rights and/or ending humanity
That 671B parameter model requiring 404GB are pure madness. These monsters are:
Seeing these on a dropdown menu is like finding a "detect dark matter" setting on your microwave. Sure, it's technically impressive, but do you really need it to heat up your leftovers?
The Actual Useful Advice Section
If you've made it this far, here's the TL;DR that should have been at the top:
Remember, a well-tuned smaller model will often outperform a generic larger one. That 7B model specifically fine-tuned for coding will write better Python than a general-purpose 70B model that's trying to be all things to all people.
In the end, the best parameter size is the one that runs on your hardware, solves your problem, and doesn't require you to take out a loan for your electricity bill.
Author's note: This article was written with the assistance of an AI that refused to specify its parameter count, simply saying it was "adequate for the task at hand."
IoT Consultant, Trainer
5 天前Like the way you articulated
Is it Dave or Sam? Both? I bet on Sam as refusing to specify its parameters is really Sam :)
CTO at Hashmeta, Co-Founder of Business+AI, AI practitioner, Web3 enthusiast, Technopreneur. Leveraging my unique experience across academia, startups, and large consulting firm to bridge technology and business.
6 天前Like the usefulness and your fun writing style!
Notary Public, Commissioner for Oaths, Advocate & Solicitor
6 天前Uli Hitzel?did you try the 671B monsters
Driving Digital Transformation & Unlocking Business Potential with Technology || Biopharma/Lifesciences || Retail/CPG/RFA
6 天前Brilliant description. Thank you for putting this together..