"70B parameters" counts every learnable value; weights dominate that count.
A neural network is a giant function with millions-to-trillions of adjustable numbers. Each one is a parameter. Most parameters are weights — multipliers on inputs. A smaller set are biases — additive shifts. Both are learned during training (Google AI Glossary; Stanford CS231n).
For a single neuron:
output = activation(w1*x1 + w2*x2 + ... + wn*xn + b)
w1 ... wn are weightsb is a biasIn a 175B-parameter model, ~98% are weights, ~1-2% are biases, and a tiny fraction are layernorm scales and other learned scalars.
| Term | Learned? | Examples |
|---|---|---|
| Weight | Yes | Connection strengths |
| Bias | Yes | Per-neuron offsets |
| Parameter | Yes | Weights + biases + other learned scalars |
| Hyperparameter | No (set before training) | Learning rate, batch size, number of layers |
The big distinction: parameters change during training; hyperparameters do not.
Is "7B parameters" the same as "7B weights"? Close enough for marketing. Technically includes a small number of non-weight parameters.
Are activations parameters? No — activations are computed at runtime, not stored or learned.
Are embeddings weights? Yes — the embedding table is a big weight matrix.
Do biases matter? A little — some modern transformers drop biases to simplify without losing much accuracy.
What is parameter efficiency? Techniques like LoRA update <1% of parameters and match full fine-tuning quality for many tasks.
How do I count parameters? sum(p.numel() for p in model.parameters()) in PyTorch.
Does more parameters mean smarter? Roughly, but diminishing returns. 70B tuned model > 175B untuned model.
Weights are the dominant type of parameter; in most sentences the two words are interchangeable. Distinguish parameters from hyperparameters to avoid confusion. More on Misar Blog.
Free newsletter
Join thousands of creators and builders. One email a week — practical AI tips, platform updates, and curated reads.
No spam · Unsubscribe anytime
A complete list of 25 free AI writing tools in 2026 — Claude, ChatGPT, Gemini, Grammarly, QuillBot, Hemingway, and more…
The top free AI image generators in 2026 — DALL-E via Bing, Gemini, Ideogram, Leonardo, Stable Diffusion, Flux — with qu…
The top free AI tools for nonprofits in 2026 — grant writing, donor outreach, social posts, translations, research — wit…
Comments
Sign in to join the conversation
No comments yet. Be the first to share your thoughts!