Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
0 points
hnuser123456
7mo ago
0 comments
Share
A 24GB GPU can run a ~30b parameter model at 4bit quantization at about 8k-12k context length before every GB of VRAM is occupied.
undefined | Better HN
0 comments
default
newest
oldest
iamnotagenius
7mo ago
Not quite true. Depends on number of KV heads. GLM4 32b at IQ4 quant and Q8 context can run full context with only 20GiB VRAM.
j
/
k
navigate · click thread line to collapse