No, MLX is nothing like a Cuda translation layer at all. It’d be more accurate to describe MLX as a NumPy translation layer; it lets you write high level code dealing with NumPy style arrays and under the hood will use a Metal GPU or CUDA GPU for execution. It doesn’t translate existing CUDA code to run on non-CUDA devices.
What do you mean when you say "run"? Low graphics 45 FPS at 720p? Or ultra graphics 120 FPS at 4k? My assumption is that a fairly large part of that space is inaccessible with the integrated GPU.
Full quality Sora yeah probably needs serious hardware. distilled version on a 4090 though? maybe. danjl earlier in this thread made a solid case for just distributing the weights and letting people run locally. the SD/ComfyUI crowd already does this daily. OpenAI won't because deepfakes. they already had a mess WITH server-side moderation. open weights with zero moderation, good luck with that PR.
The whole decomp and recomp scene has really started popping off this past year in the emulation community. The N64 recomp tool was the first one to really get interest going.
Not now. But my question is, what was stopping IBM from doing what Arm did? We are where we are now and it's too late. But as far as I can see, there was nothing too special about Arm as compared to PowerPC back then, on a technical level.
Price competitive to AMD and intel? Sure. Abilities? There is no magic, the Tellium and Power11 are each as complicated as something like Epyc and the former has both a longer and taller compatibility totem pole than x86.
Anyway this post was never about building ARM or x86 CPUs, the point is they could have done a zArch fast path for x86 for "free", so there is some other strategy at play to consider doing it with ARM.
In the era of Windows 95, having a network connection was still a rarity.
Expecting modern systems to have package management and sandboxing mechanisms would have been 20 years ahead of their time.
reply