Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.
View attachment 2376453

@eas

I'm literally dumbfounded trying to figure out how you disagreed with me here. In a recent interview with NVIDIA's CEO Jensen Huang, he made this exact statement.
There are people who are massively over invested in NVIDIA, just as there are people who are massively over invested in Apple. The NVIDIA acolytes will deny it with the same ferocity the Apple acolytes deny it.
 
  • Like
Reactions: d0sed0se
I thought Apple could have released a Mac Pro that was upgradable by adding boards/blades: processing, storage, etc. These boards would be interconnected using something like InfiniBand or a proprietary bus based on a torus topology.

The way I see it (I might be missing something here), there would be no need to come up with these “sticking two Ax Pro/Max to each other” ideas, and machines could escalate to ridiculous amounts of processing, memory and storage, for rendering videos, scientific computing, AI and other massive workloads.
I would imagine they would be trying to have multiple SOC Mx Ultra chips on Mother board/cards to support more memory and processing power. I am glad Apple is trying to use their own Servers, M2 Ultra isn’t near Nvidia, but it’s a good start to improve.
 
  • Like
Reactions: erthquake
You are missing the point, Apple, according to this article, is not making an AI specific chip, but is trying to use their outdated M2 Ultra to compete with hardware that are much faster in AI and more efficient ultimately.
It's like trying to mine crypto with GPU and compete with ASIC miners that are built specifically for that task.
If Apple makes their own optimized hardware for AI, something that Google did for Gemini's LLM and was successful with it, is a whole other story.
Apple also thought that they can compete with Qualcomm for modem chip because they were successful to make their own SoC, but failed miserably, I wouldn't claim they can catch up to Nvidia in AI anytime soon unless it actually happens, but that's Apple's Tim Cook in a nutshell these days, try to catch up with everyone else because you are always behind.
Those who believe Apple can compete with Nvidia in AI using M2 Ultra are the same people who believed 8GB of ram on mac is 16GB on windows .
Let's face it, we all know exactly why Apple is not going to use Nvidia GPU's.
You missed what I said. For now Apple can continue to use hybrid solutions such as tapping AWZ, Google cloud for training until open sourced UXL stack for GPU's is matured by year end or even longer, if needed. For B2C inferencing, a lot of it will be done on device and some on the the M2 Ultra servers. When AI GPUs sporting open source UXL stack are readied by Google, Intel, Qualcomm, ARM, etc. then they can replace their current hybrid solutions with the open source options. No need for Apple to depend directly on Cuda platform. Let the big hyperscalers do it for you as they are all collaborating with each other to replace NVIDIA anyway.

As I said earlier, there is a lot of insider selling at NVIDIA for a reason. They don't know how much longer they can keep up their first mover effect for the AI server sector. Even with the brute force of their GPU revenues, I can assure you they will not be announcing any share buy back or dividend on the scale of what Apple did last week. If they are confident of their future they would. I bet they don't when they report later this month. This is probably because they do not know where they will be in 5 years when Apple clearly does have confidence in their own future revenues.
 
Last edited:
as someone who uses chatgpt i can see what the hype is all about, its like when google and youtube came out. However I would love if apple just integrated with chapgpt instead of creating their own, its like the iPhone using google search engine and I believe the majority use google maps. Time will tell, but i still don’t use siri after all the years since it came out.
They heard you!

 
Kinda like Amazon’s stores where they claimed to use AI when instead it was just a bunch of cameras with low paid Indian workers watching them.
No way, is that really how that ended up going down? I assumed they failed because it just didn't work and/or wasn't profitable.
 
There are people who are massively over invested in NVIDIA, just as there are people who are massively over invested in Apple. The NVIDIA acolytes will deny it with the same ferocity the Apple acolytes deny it.

I'll say AAPL has proven to provide more long-term substance. NVDA may yet prove the same, but for now, AI is arguably in a hype cycle.

And some of Huang's assertions make me think he's either pretending to be clueless about LLM capabilities (and non-capabilities) and deliberately exaggerating to trick investors, or he's really chasing something that won't be possible for decades, if ever.
 
  • Like
Reactions: Zdigital2015
Seems solid to scale.

  • A14 Bionic (iPad 10): 11 Trillion operations per second (OPS)
  • A15 Bionic (iPhone SE/13/14/14 Plus, iPad mini 6): 15.8 Trillion OPS
  • M2, M2 Pro, M2 Max (iPad Air, Vision Pro, MacBook Air, Mac mini, Mac Studio): 15.8 Trillion OPS
  • A16 Bionic (iPhone 15/15 Plus): 17 Trillion OPS
  • M3, M3 Pro, M3 Max (iMac, MacBook Air, MacBook Pro): 18 Trillion OPS
  • M2 Ultra (Mac Studio, Mac Pro): 31.6 Trillion OPS
  • A17 Pro (iPhone 15 Pro/Pro Max): 35 Trillion OPS
  • M4 (iPad Pro 2024): 38 Trillion OPS
Seriously trying to say my iPhone 15 pro is more powerful than my M1 Pro MacBook Pro 14”?
 
I have a Nvidia 4090 workstation. My M1 Max runs stuff 4090 can only dream. Nvidia need to up the 24 GB VRAM or lower prices of GPU with more than 40 GB. If Apple can provide 256 GB unified memory, I can lower my cloud costs(already low with M1 Max).
Apple needs to allow multiple studios to have cluster of GPU and memory.
What can the M1 Max do a 4090 can’t?
 
Maybe just use Nvidia GPU's that are like 50 times faster and are actually built for this type of workloads ?
That’s because of the software stack making specific use of cuda cores. Apple having their own software stack to utilize its GPUs will do the same. Nvidia is about to be last in line between AMD and intel as far as price/performance goes/power usage goes. AMD’s mi300x is already better in many cases at 1/3rd the cost and the mi350x could arguably be the best out there, again cheaper, but it depends on the workload. AI is very broad terms and there’s TONS of specific calculations based on what’s being modeled. Nvidia will just pick the few they do better than. Many of the big key players are signing AMD contracts now and as the AMD software stack matures and is most importantly free and open development it’s expected to see the market share of AMD start taking from Nvidia.
 
What can the M1 Max do a 4090 can’t?
4090 runs out of memory, and doesn’t work with midsized models or higher. For those saying 4090 is gaming GPU, 2X4090 easily matched A100 performance in most inferences. H100 has lot more built in accelerators, but 4x 4090 is decent. Problem is Nvidia wouldn’t put more RAM in 4090, it will eat up higher end GPU sales. For cloud providers, they can stack up 8+ H100 GPUs and milk money. Apple has great opportunity to take the middle market.
 
You missed what I said. For now Apple can continue to use hybrid solutions such as tapping AWZ, Google cloud for training until open sourced UXL stack for GPU's is matured by year end or even longer, if needed. For B2C inferencing, a lot of it will be done on device and some on the the M2 Ultra servers. When AI GPUs sporting open source UXL stack are readied by Google, Intel, Qualcomm, ARM, etc. then they can replace their current hybrid solutions with the open source options. No need for Apple to depend directly on Cuda platform. Let the big hyperscalers do it for you as they are all collaborating with each other to replace NVIDIA anyway.

As I said earlier, there is a lot of insider selling at NVIDIA for a reason. They don't know how much longer they can keep up their first mover effect for the AI server sector. Even with the brute force of their GPU revenues, I can assure you they will not be announcing any share buy back or dividend on the scale of what Apple did last week. If they are confident of their future they would. I bet they don't when they report later this month. This is probably because they do not know where they will be in 5 years when Apple clearly does have confidence in their own future revenues.
The future of ai is local, that means on mobile devices which are the main computers that we use. Nvidia, Google and ibm are developing enterprise hardware.

Therefore apple is running its own race. Mobile use case requires making costs and efficient top priorities.

As a result I think that Apple are already developing an in-house stack from hardware, model, training to inference and evaluation. Long term hybrid approach won't be good optimal.

Nvidia hardware is currently amazing for training but overkill and inefficient for inference compared with groq. Eventually even training will be run locally on a user's own document or content.
 
  • Like
Reactions: erthquake
What can the M1 Max do a 4090 can’t?

It’s about the size of the model. Larger models take up more memory whilst processing, and when computed using the GPU all of that needs to be stored in VRAM.

Apple’s unified architecture has an advantage here because all system RAM is addressable as VRAM - up to 128GB on M1 Ultra and 192GB on M2 Ultra. And then one step further because the same memory can be used by the NPU as well at the same time.

So if you are running a model that exceeds the available VRAM on a 4090 it doesn’t matter how much power it has available because it can’t use it effectively. In such cases a Mx Ultra and even Mx Max chips will be far more performant even if they have less absolute compute power.

To get systems from other vendors with Mx Ultra levels of addressable VRAM costs an order of magnitude more than the amount Apple currently sells the entire system for (Mac Studio for example) to consumers.

From there, take off the 35% margin, and pack the chips into a rack that has a lower production cost per chip and the cost to Apple is really low. It’s a major competitive advantage if they can leverage it effectively. Also, M2 Ultra is produced on a now very mature, high yielding process so the cost per chip is only going down over time.

The point being, even if they have to use more chips to match performance, they can still save an enormous amount of money versus buying from the competition.
 
Seriously trying to say my iPhone 15 pro is more powerful than my M1 Pro MacBook Pro 14”?
We're (or at least should be) talking strictly about the NPU performance here.

The new grail is NPUs capable of 40 TOPS according to some chipmakers that are reaching that level for the next generation of AI-centric chips.
 
I hope this means Siri will soon stop being useless.
The AI hardware just means the AI functions aren't going to use as much electricity, still need the proper software to be of any use. I find it hard to believe Siri hasn't been improved because apple didn't want to use electricity.
 
  • Haha
Reactions: AgeOfSpiracles
one query. It was confirmed that the M1, M2 and M3 series all contain a security flaw within the hardware. With comments about customer security is it wise to utilise flawed hardware?

With regards the latest buzz of AI, it is worth perusing the following article as its nothing new. Even Siri with all its faults can be described as AI, as are the awful newer answerphone systems that seem to confer the title Artificial Ignorance rather than 'intelligence'

The security flaw you describe depends cyrptographic processes being run on the Performance cluster, and the hostile app needs to be run on the same cluster at the same time to have a chance of exploiting the flaw.

Even if Apple runs some cryptographic processes on these ML servers, I can't imagine they'll just run random apps from online on them, so there's very little risk of that issue being exploited.
 
you are also missing the obvious: the M2 Ultra is a niche product, and bigger chips (the so-called M1/M2 Extreme) were cancelled because of that.

Extreme was just an unfounded rumor. There are hardware-level things saying, for example, whether an interrupt comes from chip 1 or chip 2, and it doesn't support having more than 2. M1 Ultra is two M1 Max connected together, and there is no way to connect four M1 Max together.

If Extreme was planned and cancelled, it was cancelled several years before the M1 was launched.

Something that does seem weird to me is that M3 Max apparently doesn't have a way to connect two of them together (Ultrafusion). So I guess M3 Ultra is either not gonna happen or it's going to be an entirely different chip.
 
You are missing the point, Apple, according to this article, is not making an AI specific chip, but is trying to use their outdated M2 Ultra to compete with hardware that are much faster in AI and more efficient ultimately.
It's like trying to mine crypto with GPU and compete with ASIC miners that are built specifically for that task.
If Apple makes their own optimized hardware for AI, something that Google did for Gemini's LLM and was successful with it, is a whole other story.
Apple also thought that they can compete with Qualcomm for modem chip because they were successful to make their own SoC, but failed miserably, I wouldn't claim they can catch up to Nvidia in AI anytime soon unless it actually happens, but that's Apple's Tim Cook in a nutshell these days, try to catch up with everyone else because you are always behind.
Those who believe Apple can compete with Nvidia in AI using M2 Ultra are the same people who believed 8GB of ram on mac is 16GB on windows .
Let's face it, we all know exactly why Apple is not going to use Nvidia GPU's.
View attachment 2376505
The yield on older chips like M2 is probably quite high, and the cost much lower than M3/4 chips. Apple can probably just throw a lot of cheap chips at the problem until they learn enough to design hardware better optimized for inference (and training?). Remember, Apple's not competing with Nvidia to sell their AI server hardware. They're just making a cost decision solving their own compute needs.
 
  • Like
Reactions: spittt
as someone who uses chatgpt i can see what the hype is all about, its like when google and youtube came out. However I would love if apple just integrated with chapgpt instead of creating their own, its like the iPhone using google search engine and I believe the majority use google maps. Time will tell, but i still don’t use siri after all the years since it came out.
Long term, I want Apple's commitment to privacy. ChatGPT now has "memory," but I'd rather that information live on my iPhone's Secure Enclave like my health data.
 
The Secure Enclave doesn't contain health data. Think of it more as a keychain. The health data gets encrypted and stored on regular flash, but significant, essential portions of the encryption key reside only on the Secure Enclave.
 
  • Like
Reactions: erthquake
The Secure Enclave doesn't contain health data. Think of it more as a keychain. The health data gets encrypted and stored on regular flash, but significant, essential portions of the encryption key reside only on the Secure Enclave.
Ah. Thanks for that clarification.
 
  • Like
Reactions: chucker23n1
I thought that it's feasible for Apple to release the M4 Ultra for the Mac Studio and/or Mac Pro this summer, but we got now two different rumours saying that the AI servers will have the M2 Ultra for the moment and will update to M4 later, which seems to suggest that maybe we are not going to get any M4 Ultra this summer.
 
Extreme was just an unfounded rumor. There are hardware-level things saying, for example, whether an interrupt comes from chip 1 or chip 2, and it doesn't support having more than 2. M1 Ultra is two M1 Max connected together, and there is no way to connect four M1 Max together.

If Extreme was planned and cancelled, it was cancelled several years before the M1 was launched.

Something that does seem weird to me is that M3 Max apparently doesn't have a way to connect two of them together (Ultrafusion). So I guess M3 Ultra is either not gonna happen or it's going to be an entirely different chip.
yes, my understanding of any bigger chips, there was no reason to move past design phase as any market fit study showed there just wont be enough sales to justify creating one. apple silicon servers changes that, even if they are purely internal products.
by the very fact of how fast Apple went from M3 to M4 we might assume the M3 was just a stop-gap generation and if Ultra happens again, it will be based on M4.
only question is, how fast TSMC is able to ramp up production/yield to release bigger M4 chips (and just to cool down any expectations, the they need to account for A18 production too)
 
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.