A good technical project, but honestly useless in like 90% of scenarios.
You want to use an NVidia GPU for LLM ? just buy a basic PC on second hand (the GPU is the primary cost anyway), you want to use Mac for good amount of VRAM ? Buy a Mac.
With this proposed solution you have an half-backed system, the GPU is limited by the Thunderbolt port and you donât have access to all of NVidia tool and library, and on other hand you have a system who doesnât have the integration of native solution like MLX and a risk of breakage in future macOS update.
I don't know how Apple has evaded regulatory scrutiny for their refusal to sign Nvidia's eGPU drivers since 2018.
mlfreemanyesterday at 7:04 PM
I followed the instructions link and read the scripts...although the TinyGPU app is not in source form on GitHub, this looks to me like the GPU is passed into the Linux VM underneath to use the real driver and then somehow passed back out to the Mac (which might be what the TinyGrad team actually got approved).
Or I could have totally misunderstood the role of Docker in this.
Keyframeyesterday at 6:33 PM
Such a shame both companies are big on vanity to make great things happen. Imagine where you could run Mac hardware with nvidia on linux. It's all there, and closed walls are what's not allowing it to happen. That's what we as customers lose when we forego control of what we purchase to those that sold us the goods.
tensor-fusionyesterday at 7:16 PM
As more people carry ARM laptops and keep the GPU somewhere else, I think the interesting UX question becomes whether the GPU can "follow" the local workflow instead of forcing the whole workflow to move to the GPU host. That's the problem we've been looking at with GPUGo / TensorFusion: local-first dev flow, remote GPU access when needed. Curious whether people here mostly want true attached-eGPU semantics, or just the lowest-friction way to access remote compute from a Mac without turning everything into a remote desktop / VM workflow.
arjieyesterday at 5:14 PM
Woah, this is exciting. I'm traveling but I have a 5090 lying around at home. I'm eager to give it a go. Docs are here: https://docs.tinygrad.org/tinygpu/
I hope it'll work on an M4 Mac Mini. Does anyone know what hardware to get? You'll need a full ATX PSU to supply power, right? And then tinygrad can do LLM inference on it?
lowbloodsugartoday at 1:17 AM
Can I do prefill on the eGPU and the decode on the Mac?
vonduryesterday at 6:52 PM
If you could get Nvidia driver support on Macâs I bet Apple would have sold more MacProâs.
eoskxyesterday at 5:13 PM
Interesting, but cannot run CUDA or more to the point `nvidia-smi`.
dd_xploreyesterday at 6:55 PM
Why does Apple need to make the drivers in a walled garden? Atleast they should support major device categories with official drivers.
wmfyesterday at 4:59 PM
Pretty misleading. This driver is only for compute not graphics.
the__alchemistyesterday at 6:21 PM
I'm writing scientific software that has components (molecular dynamics) that are much faster on GPU. I'm using CUDA only, as it's the eaisiest to code for. I'd assumed this meant no-go on ARM Macs. Does this news make that false?
frankcyesterday at 5:45 PM
My main thought is would this allow me to speed up prompt process for large MoE models? That is the real bottleneck for m3ultra. The tokens per second is pretty good.
What are the limitations of USB4/Thunderbolt compared with a regular PCIe slot?
userbinatoryesterday at 8:36 PM
[flagged]
vegabookyesterday at 8:56 PM
now can they please approve the linux kernel
bigyabaiyesterday at 4:57 PM
The opportunity cost of Apple refusing to sign Nvidia's OEM AArch64 drivers is probably reaching the trillion-dollar mark, now that Nvidia and ARM have their own server hardware.