Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
Show HN: Selfhostllm.org – Plan GPU capacity for self-hosting LLMs
(opens in new tab)
(selfhostllm.org)
7 points
erans
7mo ago
3 comments
Share
A simple calculator that estimates how many concurrent requests your GPU can handle for a given LLM, with shareable results.
Show HN: Selfhostllm.org – Plan GPU capacity for self-hosting LLMs | Better HN
3 comments
default
newest
oldest
erans
OP
7mo ago
I also added a Mac version:
https://selfhostllm.org/mac/
so you can know which models you can run on your Mac and get an estimated tokens/sec.
harshnigam
7mo ago
I see it doesn't take GPU performance into consideration when showing the estimates. H100 and A100 are performing the same. Am I doing it wrong?
atmanactive
7mo ago
Very useful, thanks. I'm missing a reset button though.
j
/
k
navigate · click thread line to collapse