Darajavahus
macrumors regular
Jesus if the Pro have the same CPU, the multi is 2x my M3 Pro ...just two years later... ðŸ˜
That would likely be well into the five digits, with current RAM prices.Really hope they come out with a M5 Ultra with 1TB+ memory Studio.. literally everyone wants to run local LLMs. Come on Apple!
Hope so. But they discontinued even the 512GB M3 Ultra now. Only a 256GB config remains…Really hope they come out with a M5 Ultra with 1TB+ memory Studio.. literally everyone wants to run local LLMs. Come on Apple!
Any idea what 1TB of ram cost on the 3rd party market today? Roughly $30k. I can’t imagine how expensive that would be in Apple $$$Really hope they come out with a M5 Ultra with 1TB+ memory Studio.. literally everyone wants to run local LLMs. Come on Apple!
Going from 12 performance cores to 6 (yes, SIX "super cores" which ARE equivalent to the old performance cores) is quite an enormous drop.The improvement over M4 max is only 14%? That’s not a lot. In the worst intel days the improvements are around 5%
With identical 18 CPU cores, M5 Pro will likely score the same (synthetic benchmarks probably don't reflect Max's faster RAM). Insane.Jesus if the Pro have the same CPU, the multi is 2x my M3 Pro ...just two years later... ðŸ˜
How are you managing to get by without any neural accelerators or ray tracing? 😉I'm waiting for the M6 series of processors. The M5 Pro and Max are the first set of processors using an "all-new Fusion Architecture". I'm interested in the refinements that come in the updated version. That and my M1 MacBook Pro is still fast without any lag or problems.
Would love to see some LLM benchmarks to see if that 4x performance claim over the M4 Max for AI tasks is accurate
You're missing the new tensor cores in the M5 series.No need to guess. With how current LLM work, memory bandwidth basically determines the speed. If the bandwidth isn’t 4× higher, the speed won’t be 4× either. It’s just physics.
Really hope they come out with a M5 Ultra with 1TB+ memory Studio.. literally everyone wants to run local LLMs. Come on Apple!
You're missing the new tensor cores in the M5 series.
I am an M1 16 Max waiter as well. Very excited by the improvements, especially internal disk speeds and the TB5 improvements, but I just know I would have big time buyers remorse if I didn’t wait for the redesign and the new panels. Was a twitch away from ordering a Max 40c with 4TB and 128GB, now just praying that they don’t cripple hammer anything in the M6 transition and that the price hike isn’t insane.I have M1 Pro 16", Im waiting for the M6 if thats when the redesign is; I dont need that much power so as good as this sounds to upgrade, not really worth it in my personal situation. Love all the improvements adding up.
How do I put this gently... This comment is very "2 years ago." A lot has changed. Too much to explain in a comment. The current SOTA open weight models perform extremely well - and they need this amount of memory to run at or close to trained weights.I’m actually more curious what people expect to run with 1TB for local LLM.
Even the biggest open-source model we have right now, like Llama 3.1 405B, would be around ~800GB if you ran it unquantized in FP16. But realistically almost nobody runs local models like that anymore, quantization is basically the norm now.
With something like Q4 / int4, even a 405B model would only need roughly ~200GB just for the weights.
And with MoE models, the parameter count can look huge on paper, but during inference only a subset of experts are actually active anyway, so the real runtime footprint is way smaller than the headline numbers.
Also another thing people forget: even if someone released an open model with parameter counts somewhere near GPT-4 scale, that still doesn’t mean the real world performance would match it. Model capability depends on way more than just parameter count : training data, post-training, alignment, architecture, all that stuff matters a lot.
And that’s even assuming the model would actually need anywhere close to 1TB of memory in the first place.
I have no interest in running LLMs, therefore, not everyone wants to run LLMs.Really hope they come out with a M5 Ultra with 1TB+ memory Studio.. literally everyone wants to run local LLMs. Come on Apple!
How do I put this gently... This comment is very "2 years ago." A lot has changed. Too much to explain in a comment. The current SOTA open weight models perform extremely well - and they need this amount of memory to run at or close to trained weights.
Also the Max has double the bandwidth and double media encoders. So if AI or encoding is important, Max is the preferred choice.Hang on. It was written:
"With four additional CPU cores compared to the M4 Pro, the new CPU architecture significantly boosts multithreaded performance by up to 30%."
"The M5 Max chip pairs the 18-core CPU [...]. The new CPU architecture offers up to 15% higher multithreaded performance when compared to the M4 Max."
Seems strange. But then:
M4 Pro = 10 P-cores, 4 E-cores => M5 Pro = 6 S-cores, 12 P-cores
M4 Max = 12 P-cores, 4 E-cores => M5 Max = 6 S-cores, 12 P-cores
This means the M5 Pro is equal to the M5 Max in CPU cores. And the better buy from the CPU core perspective?
$3,399 14" M5 Pro MBP 64GB RAM 2TB SSD (20 GPU cores)
$4,299 14" M5 Max MBP 64GB RAM 2TB SSD (40 GPU cores)
$900 difference for 20 GPU cores
Thanks! Then, looking at the CPU performance only, I wonder how much double the memory bandwidth affects the benchmarks.Also the Max has double the bandwidth and double media encoders. So if AI or encoding is important, Max is the preferred choice.
The M1-M2 Pros/Maxes had the same CPU core count. In addition to double the GPU cores, you also get double the memory bandwidth, a second media encoder, double the number of displays it can drive. Probably still not worth $900 but it's more than just the GPU cores.Hang on. It was written:
"With four additional CPU cores compared to the M4 Pro, the new CPU architecture significantly boosts multithreaded performance by up to 30%."
"The M5 Max chip pairs the 18-core CPU [...]. The new CPU architecture offers up to 15% higher multithreaded performance when compared to the M4 Max."
Seems strange. But then:
M4 Pro = 10 P-cores, 4 E-cores => M5 Pro = 6 S-cores, 12 P-cores
M4 Max = 12 P-cores, 4 E-cores => M5 Max = 6 S-cores, 12 P-cores
This means the M5 Pro is equal to the M5 Max in CPU cores. And the better buy from the CPU core perspective?
$3,399 14" M5 Pro MBP 64GB RAM 2TB SSD (20 GPU cores)
$4,299 14" M5 Max MBP 64GB RAM 2TB SSD (40 GPU cores)
$900 difference for 20 GPU cores