Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.
Is this some kind of joke? MS spends way more for their AI server infrastructure. You're acting as if $4.5 billion is a lot for Apple to spend given their total user base. $4.5 billion for a company the size of Apple is table stakes.

Oh wait I was given an outdated article from 2018. Never mind I'm erasing that last bit
 


Apple is expected to spend several billion on hardware to support its artificial intelligence development in 2024, according to speculation from Apple analyst Ming-Chi Kuo. Kuo expects Apple to spend "at least" $620 million on servers in 2023 and $4.75 billion on servers in 2024.

hey-siri-banner-apple.jpg

Apple could purchase between 2,000 and 3,000 servers this year, and up to 20,000 next year. Kuo thinks that Apple is purchasing servers equipped with Nvidia's HGX H100 8-GPU for generative AI training, with the company planning to upgrade to B100 next year. Nvidia calls its H100 an AI supercomputing platform, and each one is priced at around $250,000.

Kuo appears to be guessing at Apple's purchasing plans here, and he says that he expects Apple will use AI servers it is purchasing and installing itself to train large language models rather than virtual hosting from other cloud service providers for improved security and privacy. He does say that Apple could develop its own server chips to save on server costs, but he has seen no evidence that Apple is doing that at this time.

While Apple appears to be making a major investment into AI, Apple's server purchasing will fall behind other companies like Meta and Microsoft. Apple will also need to invest in labor costs, infrastructure, and more, and Kuo suggests that Apple will need to spend several billion dollars annually to have a chance of catching up with competitors. Kuo claims that he is "genuinely concerned" about the future of Apple's generative AI business if Apple spends just a billion dollars a year as suggested by Bloomberg's Mark Gurman.

Over the weekend, Gurman said that Apple is on track to spend $1 billion per year on its AI efforts. Gurman says that Apple is working on a new, smarter version of Siri and is aiming to integrate AI into many Apple apps.

Article Link: Kuo: Apple Could Spend $4.75 Billion on AI Servers in 2024
I think Kuo needs to understand how much $1 billion dollars US is, or maybe he's talking about HK$ & forgot to put the HK in front 🙄
 
  • Disagree
Reactions: Anthony20022
Better hope China does not invade Taiwan - where do you think Nvidia fabricates all these chips at? With the AI-boom, TSMC is even more a critical supplier than before. I'm glad to see several chip fabs are being built in the US again.. but it will take years to rival TSMC's super fabs sheer scale and wafers per day output.
 
  • Like
Reactions: nt5672
Microsoft doesn't have an ecosystem lock compared to Apple.

Apple Music, TV, News, Maps, iMessage, iPhone, Mac, etc. All of those working together using one AI is worth spending on.
But this articles say they are spending Billions on the hardware alone. That does not include the cost to port all those applications to use the new information source from the hardware. Those are 2 very different things. If Apple spends 4.8 Billion to buy the hardware...it will cost another billion maybe in development costs to make it all work with the applications you listed. This is a very delicate and time consumer effort. There is testing...then dev...then QA then UAT.
 


Apple is expected to spend several billion on hardware to support its artificial intelligence development in 2024, according to speculation from Apple analyst Ming-Chi Kuo. Kuo expects Apple to spend "at least" $620 million on servers in 2023 and $4.75 billion on servers in 2024.

hey-siri-banner-apple.jpg

Apple could purchase between 2,000 and 3,000 servers this year, and up to 20,000 next year. Kuo thinks that Apple is purchasing servers equipped with Nvidia's HGX H100 8-GPU for generative AI training, with the company planning to upgrade to B100 next year. Nvidia calls its H100 an AI supercomputing platform, and each one is priced at around $250,000.

Kuo appears to be guessing at Apple's purchasing plans here, and he says that he expects Apple will use AI servers it is purchasing and installing itself to train large language models rather than virtual hosting from other cloud service providers for improved security and privacy. He does say that Apple could develop its own server chips to save on server costs, but he has seen no evidence that Apple is doing that at this time.

While Apple appears to be making a major investment into AI, Apple's server purchasing will fall behind other companies like Meta and Microsoft. Apple will also need to invest in labor costs, infrastructure, and more, and Kuo suggests that Apple will need to spend several billion dollars annually to have a chance of catching up with competitors. Kuo claims that he is "genuinely concerned" about the future of Apple's generative AI business if Apple spends just a billion dollars a year as suggested by Bloomberg's Mark Gurman.

Over the weekend, Gurman said that Apple is on track to spend $1 billion per year on its AI efforts. Gurman says that Apple is working on a new, smarter version of Siri and is aiming to integrate AI into many Apple apps.

Article Link: Kuo: Apple Could Spend $4.75 Billion on AI Servers in 2024
It's a load of rubbish speculation for clicks.
What's really amusing is that most people who are desperate for AI, don't realise it's going to replace them 🤣
 
  • Like
Reactions: nt5672
But this articles say they are spending Billions on the hardware alone. That does not include the cost to port all those applications to use the new information source from the hardware. Those are 2 very different things. If Apple spends 4.8 Billion to buy the hardware...it will cost another billion maybe in development costs to make it all work with the applications you listed. This is a very delicate and time consumer effort. There is testing...then dev...then QA then UAT.

It comes down to the value of the tie-in. Apple is all about their ecosystem and if their AI glues everything together, that's their main selling point.

Apple is already spending $1B on car with nothing to show for it.
 
  • Like
Reactions: klasma and jamezr
Microsoft will spend 25 billion.

Meta will spend more them 10 billion.

If apple manages to bring AI to the masses, they will prevail once again.
 
Apple has been softening on it's Nvidia stance for the last year or so. One of the interviews when the Mac Pro was updated for Apple Silicon literally praised Nvidia as the best platform for creating AI models. I wouldn't be surprised if we see Nvidia return to the Mac Pro for specialty compute and AI workloads. The bad blood between Apple and Nvidia may finally be dissolving. I wouldn't be surprised if we see Apple produce Nvidia GPUs on their reserved TSMC capacity if there are supply chain constraints.

One thing to note is that Microsoft hosts a service to build AI models in the cloud. Apple wouldn't need nearly as many as Microsoft since they would be unlikely to offer it as a service for others.
 
Last edited:
  • Like
Reactions: nt5672
Yes, but that's also an opportunity perfectly suited for Apple... as they can optimize the training models to be smaller and more efficient to process. I've appreciated that Apple keeps their resources (RAM, etc.) lighter... as it forces developers to work harder to optimize their apps. As soon as you give developers massive amounts of headroom, they get lazy, fast.
That’s what everyone is working on, because less resource consumption = more profit. It’s rather difficult though.
 
  • Like
Reactions: CarAnalogy
Top end NVIDIA DGX H100 servers cost $500k each. A billion dollars doesn't go far and I think Kuo's prediction is closer to the mark.


And I bet Apple is absolutely thrilled having to pay them that much money after their long drawn out spats over GPUs in Macs and many other things.
 
  • Like
Reactions: JPack
Apple is more likely to invest in AMD EPYC Servers and the very soon to be released MI300, over Nvidia.

AMD has basically no development platform for AI. NVIDIA has a 5+ year head start (conservative estimate) with CUDA.

Apple isn't going to reinvent the wheel when time is key. It would be like using RISC-V instead of ARM.
 
Yes, but that's also an opportunity perfectly suited for Apple... as they can optimize the training models to be smaller and more efficient to process. I've appreciated that Apple keeps their resources (RAM, etc.) lighter... as it forces developers to work harder to optimize their apps. As soon as you give developers massive amounts of headroom, they get lazy, fast.

That’s one way to look at it. Very optimistic. Not untrue but I’m not sure how much it helps. That was the Unix philosophy, keep it tight and focused in every way. I hope that still endures but it’s definitely being discouraged by modern software companies.

How far we have come in terms of how much power it takes to do the same thing. Or should I say how far we’ve fallen. From simple text to graphics to video and now ten times the electricity consumption to do the same things.
 
But they do have an ecosystem lock though. Microsoft 365 is that lock. The seamless integration between OneDrive, your Office apps, Teams, Outlook, and even Windows so your projects are always with you.
100% This. MS365 and Azure are MS' fastest growing business units.
 
  • Like
Reactions: Patchwork
Imagine leveraging all those binned M1/M2 SoCs that don't have enough RAM or CPU cores to make it into a Mac or iPhone. Stick them on a custom logic board that supports dozens at once. Network them together and you could have tens of thousands of them in a rack and stay within a reasonable power and cooling envelope.

These would be extremely affordable, power-efficient, and capable.
 
Last edited:
Nah, not quite. Siri will SOUND like it's making sense and give you a very eloquent, detailed answer which will be completely wrong in every way.

Why do people trust generative AI? It's the latest stupid fad in the tech industry like NFTs were. It's going to make our current torrent of misinformation worse and cause a lot more harm than good. Just, no.
Have you actually used one? It doesn't sound like you're very familiar since you said generative which is pretty broad instead of LLM which is what this will most likely be using if it will be similar to chatGPT. They are getting better every month and they are not going away. Generative also encompasses image and audio generation, and I'm not sure what there is to not trust with that? You tell it what you want it to make and it makes it.

For me, chatGPT has really sped up my workflow as I can have it usually write 3/4 of a script and then I can finish the rest. For easy scripts it can just do the whole thing. I check over the code and test it. This ranges from something like a performant find script to run across all sites on our server, to cracking out a WordPress plugin in record time. Sometimes it makes up a weird answer, but I notice that happening less frequently, and I've been correcting it as I go. It also depends on how you talk to it. It's strange, but it's kind of like a way of programming using language. If I frame things within scopes and explain what my end goal is clearly, it tends to do better. What's also odd is I you tell it something like "take a deep breath and try again" it refocuses and often comes up with correct responses. Very strange, and we're still figuring it all out. It's the Wild West right now!

There is no comparison to NFTs, which had zero value whatsoever. I was against NFTs from the start. But this? This is extremely valuable. In the future, something like this could replace an operating system. Not joking. You could literally have it spin up an OS and customize it on the fly based on your preferences and have it connect to other services online to do whatever you want. The future is gonna be absolutely wild. We're transitioning from fixed code to code that is reactive and flexible and alive and responds to our every desire. It's an extremely exciting time to be alive! This is a new era of computing, bigger than the invention one the internet, and will have a much larger impact on society. It is up to us to understand it, and help shape and steer it and laws/regulations to make sure that it is controlled and does more good than harm. It's scary but lately I've been leaning towards being more excited than scared.

Every new major breakthrough in tech can be scary, especially to older people, and I'm nearly including myself in that category as I'm in my late 30s. But we have to realize that these are extremely early days and the best is yet to come. Things we can barely imagine will be possible in a year or two, and things that we can't even comprehend will be possible 5-10 years from now. What we are entering is the most rapid development period in all of human history by orders of magnitude. The things that we will be able to put this to use to discover are unimaginable. We're already seeing some of the early applications in science. The ability to ingest huge amounts of data and see connections and patterns that researchers miss is amazing. Everything from materials science, to discovering new types of drugs, to figuring out how to make stable fusion reactors, to a mass increase in exoplanet discovery and potential life in other systems.
 
That’s what everyone is working on, because less resource consumption = more profit. It’s rather difficult though.
Smaller models also execute faster. I'm sure everyone is trying for that, however there are probably still differences when optimizing to run on devices vs running in the cloud. Apple is likely the only one to really push running on device.
 
AMD has basically no development platform for AI. NVIDIA has a 5+ year head start (conservative estimate) with CUDA.
Which is why AMD is buying Nod.ai, so they will have a software developer ecosystem.

 
Imagine leveraging all those binned M1/M2 SoCs that don't have enough RAM or CPU cores to make it into a Mac or iPhone. Stick them on a custom logic board that supports dozens at once. Network them together and you could have tens of thousands of them in a rack and stay within a reasonable power and cooling envelope.

These would be extremely affordable, power-efficient, and capable.

It wouldn't be capable at all. AI training is massively parallel, something only GPUs offer.

M1/M2 basically has such a low level of graphics capability, it's around 100x slower compared to H100.
 
dat's a buncha green power fer all dose servers

You saw Mother Nature giving Timmy's crew a hard time in their Apple event promo. She will he hopping mad if those servers aren't driven by good old Californian sunshine.
 
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.