Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
Pool spare GPU capacity to run LLMs at larger scale
(opens in new tab)
(github.com)
11 points
i386
2d ago
3 comments
Share
Pool spare GPU capacity to run LLMs at larger scale | Better HN
3 comments
default
newest
oldest
lostmsu
2d ago
> MoE models via expert sharding with zero cross-node inference traffic
This makes the whole project questionable
vagrantJin
2d ago
This is very promising, definitely looks more user friendly than exo. Can't wait to try it out.
iwinux
2d ago
You lost me on "spare GPU". I don't have any capable GPUs, let alone spare ones :)
j
/
k
navigate · click thread line to collapse