It’s irrelevant to their core market; should they make robots and cars as well?
They tried to make cars and are apparently working on robots now
It’s irrelevant to their core market; should they make robots and cars as well?
Yeah I don’t care about that market either
There are many things you can do with a computer.
Nvidia dropped the ball on consumer level GPU
I’m very curious what you’re doing. I have a Linux workstation with a 16 core AMD and 5090. I have barely touched my Mac since the purchase. With very big LLMs I can’t keep them all in VRAM, but 32GB is still plenty big for some reasonable models.Meh. If I need to run my m4 max 128 GB RAM GPU work loads on Nvidia I need to spend 50k. Nvidia dropped the ball on consumer level GPU with 32 GB VRAM, it’s glorified junk for my purposes. If you game or can get by with low vram, buy RTX 50XX. My Linux workstation with AMD threadripper and RTX 4090 gets less and less work, I rather run 20 mins extra on m4 max than run out of GPU memory on my Linux workstation with 128 GB RAM.
What use is speed if Nvidia GPU doesn’t have enough memory, it will run out of memory and errors out. 24-32 GB GPU isn’t much, the data sets these days are much larger. Sure I can spend $$$ on cloud or just wait 20 more minute for my M4 max to finish. Either you have enough GPU memory or you don’t.Obviously depends on what you need a computer to do
What are you doing exactly?
Trying to think of a workload of any application that works out cheaper or faster in macOS on Apple silicon…..
Why stop at that, you could run 6 4090x. Good look maintaining and powering that thing. I need to pretty much revamp my whole workstation to be compatible with RTX 6000 Pro. I may get 192 GB but I can run 100 GB swap and can easily scale to around 210 GB on M4 max in case I need larger memory for a spike or shorter time. I will probably consider spark when I need to replace my workstation. I may also consider and ultra with 512 GB.You are clearly not looking at compute speed, but need the additional memory. I'm curious how you end up at 50k though. Mind to do a quick run through on components and price? (although I don't expect this to happen).
You could run four 5090 or two RTX Pro 6000 (192GB) in a single system and have the memory you need and be nowhere near 50k. You don't need the processing power of a A100/H100/etc. card as those two options are already faster when it comes to compute than M4 Max.
Here's a crazy idea, get a Spark. It's based on Grace architecture and has 128GB of unified memory. Asus is offering this (Ascent) at a cheaper price than the original Spark from Nvidia. You're looking at $3k for a single unit and can use 2x QSFP to connect two of them.
I don’t run lot of LLM. In fact if I wouldn’t spend thousands of $$$ to buy laptops/workstations to run LLM. Cloud costs or subscriptions are much cheaper to run LLM. I have large data sets I work with, and M4 Max gives me ability to lower cloud costs.I’m very curious what you’re doing. I have a Linux workstation with a 16 core AMD and 5090. I have barely touched my Mac since the purchase. With very big LLMs I can’t keep them all in VRAM, but 32GB is still plenty big for some reasonable models.
I can wait on my workstation but most likely my choice would be Spark vs Ultra studio.Aye; I was seriously contemplating purchasing the 6000 Blackwell Max-Q yesterday, and then decided that the better choice was to wait for my DGX Spark (and maybe also the M4/5 Ultra Studio!) to arrive.
Sometimes Functionality supersedes Performance.
And then there's the AMD Ryzen AI Max+ 395 which goes up to 128GB of unified memory. How does its iGPU stack up against Apple's?Here's a crazy idea, get a Spark. It's based on Grace architecture and has 128GB of unified memory. Asus is offering this (Ascent) at a cheaper price than the original Spark from Nvidia. You're looking at $3k for a single unit and can use 2x QSFP to connect two of them.
And then there's the AMD Ryzen AI Max+ 395 which goes up to 128GB of unified memory. How does its iGPU stack up against Apple's?
While I agree that Apple Silicon was designed primarily for the mobile market. However, Apple markets Apple Silicon as a viable desktop solution. We can't say out of one side of our mouth that its a mobile chip, so don't compare it to nvidia desktop gpus, and at the same time promote how awesome their desktop computers are (compared to PC desktops)I do get the laptop vs. desktop discussion, at the same time it's important to keep in mind that these are ultimately mobile chips. That's a business decision, not a technical limitation.
Apple themselves stopped comparing their GPU's to nvidia after they (mistakenly IMO) compared the M1 Ultra to the 3090 (thus causing this perpetual comparison chain).I must say am a bit confused about all the 5090 vs. M4 Max comparisons.
The 5090 RTX mobile (~ 10K shader cores, 160 watts) scores ~52K in 3DMark Wild Life Extreme
The M4 Max (~ 5K shader cores, around 70-80 watts) scores ~34K in 3DMark Wild Life Extreme
I'd it's about as expected? The other question is whether it makes sense for Apple to use >100W GPUs in their laptops.
I do get the laptop vs. desktop discussion, at the same time it's important to keep in mind that these are ultimately mobile chips. That's a business decision, not a technical limitation.
While I agree that Apple Silicon was designed primarily for the mobile market. However, Apple markets Apple Silicon as a viable desktop solution. We can't say out of one side of our mouth that its a mobile chip, so don't compare it to nvidia desktop gpus, and at the same time promote how awesome their desktop computers are (compared to PC desktops)
Apple sells the Mac Pro as a high end workstation class solution, that can easily exceed 10,000 dollars, its only natural to see how Apple Silicon matches up to Intel/AMD computationally and Nvidia for the GPU. Likewise for the Studio, its another high end desktop solution that many people are using, instead of buying a PC desktop. Apple is marketing the Mini, Studio, Mac Pro (and even the iMac) as desktop computer solutions that are in direct competition to PC desktops. Comparing desktop CPU and GPUs is warranted because that's what apple is doing
Can Apple double the shader cores and still only use 70-80 watts? How much of a die shrink would be required for this to happen?
Could they boost the clock to make up the 20K deficit and still stay around the same power usage? How much power would be required to boost the GPU clock to make up the difference?
My Studio is the best Mac and desktop I've owned, it harkens back to the days when Apple made great desktop computers at an affordable price.Studio etc. seem quite successful in their segment
From the looks of it the 5090M and the M4 Max run at about the same clock rate; the official boost for 5090M is lower by like 300Mhz, but that is assuming you are capping TGP to 95W, which no one does. Without that 65W headroom, I imagine the GPU's are a lot closer.I'd say that their architecture still has room to grow. Not sure about doubling the performance, but I think that 50% general-purposehigher performance with comparable power and die size (on the current process) should be achievable. They still have some tricks up their sleeves, for example there was a recent patent describing how to fetch data from GPU registers with lower power consumption.
For ML workloads, 4x improvement or higher should be doable at the current power consumption.
It would really surprise me if they go that route. Who knows, maybe their upcoming tech will be more flexible...
And then there's the AMD Ryzen AI Max+ 395 which goes up to 128GB of unified memory. How does its iGPU stack up against Apple's?
While I agree that Apple Silicon was designed primarily for the mobile market. However, Apple markets Apple Silicon as a viable desktop solution. We can't say out of one side of our mouth that its a mobile chip, so don't compare it to nvidia desktop gpus, and at the same time promote how awesome their desktop computers are (compared to PC desktops)
The two are not mutually exclusive. The gpu performance is most definitely “not there”
One of the few games I regularly play is “the long dark”. When I still had my last hackintosh with an rx6800 it had no problem running it at 60fps in macOS at 4k with all graphics settings at their highest
On my m4 pro it struggles to stay at 40fps with the graphics settings dialed down
Yes, that's specifically why i mentioned cars.Apple tried to make cars. Did you forget that?
What use is speed if Nvidia GPU doesn’t have enough memory,
Its fairly competitive. As it should be, on similar process tech, and AMD having 30 years of GPU development including their acquisition of ATi.
I think i understand WHY apple is doing what they're doing.
Their core market is laptops. Those running Apple desktops are a rounding error really, especially PRO users (a rounding error on the rounding error). They exist, and it sucks for them that Apple haven't really pushed the envelope there, but its a lot of R&D for a limited ROI. Most of their desktop users just don't care and are fine with even M series base.
Apple building more powerful general end user desktop GPUs would also segment the market. Part of the reason consoles work so well is that there's a few fixed configurations to optimize for. Apple isn't quite THERE but they're much closer than the PC world. And software support wise that's good.
So yeah, Apple Silicon is a mobile/laptop chip with the single exception being the Ultra, which really appears to be a casual side project for Apple unfortunately.
Doesn’t matter, it’s not all about speed. Memory matters, and it is binary.We still don’t know what you are doing so there’s no way to answer that