Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
undefined | Better HN
0 points
adastra22
3mo ago
0 comments
Share
Again, memory bandwidth is pretty much all that matters here. During inference or training the CUDA cores of retail GPUs are like 15% utilized.
0 comments
default
newest
oldest
my123
3mo ago
Not for prompt processing. Current Macs are really not great at long contexts
j
/
k
navigate · click thread line to collapse