Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

You'd have to stack 16 of these to get 2TB of VRAM, equivalent to 4 Mac Studios 512GBs chained together.

16 compared to 4. Surely even much faster networking in the Spark would degrade with that many devices?

Biggest problem with Macs is that they don't have dedicated tensor cores in the GPU which makes prompt processing very slow compared to Nvidia and AMD.



n.b. there's been a little speculation that Apple adding TensorOps to Metal 4 suggests M5/M6 may get tensor cores.

https://x.com/liuliu/status/1932158994698932505

https://developer.apple.com/metal/Metal-Shading-Language-Spe...


Nice. I hope so. That would make Macs the best local LLM machines for the masses by far.


It’s $12k for each Mac Studio, and the networking makes them only effective individually (it’s like less that 15 tokens/s with EXO) while NVLINK is very effective. The Spark is definitely more scalable, but the MLX and metal teams are cooking, so honestly either way is still winning.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: