Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.
The most critical line in this article. The "fees" prompted Apple to shop elsewhere. Thats why we wont be getting great things from Apple anymore.

Maybe they could load OpenAI models on their iPhone making the calculations on the phone an so bypassing the patent, just like with Masismo.
 
  • Like
Reactions: User 6502
The most critical line in this article. The "fees" prompted Apple to shop elsewhere. Thats why we wont be getting great things from Apple anymore.

Maybe they could load OpenAI models on their iPhone making the calculations on the phone an so bypassing the patent, just like with Masismo.

huh ... imagine that .. being able to shop around for lower fees.

(Devs dreaming about 3rd party App Stores on iOS)
 
  • Like
Reactions: bluecoast
Please don't talk about things you know nothing about.

1. Why is Apple working with Google in this space? Because Apple trains its (MULTIPLE) models on TPUs.

2. Apple has at least two known "sexy" models, along with multiple more specialized models. The two sexy models are called AFM (Apple Foundation Model). They are both *combined* language and vision models.
The on-device model is ~3B parameters, quantized to 2b per parameter.
The server model runs on Apple HW (the GPU of what are presumably M2 or M3 Ultras) and has about 200B models, quantized to about 3.5b per parameter.
Both are competitive (not obviously worse or obviously better) than counterparts *at this size* (which is definitely not the largest size, guesses are eg chatGPT5 is ~8 times larger).

Unusual (or at least non-obvious) aspects of these models include
- combined text and vision rather than two separate models
- multilingual, handling about 15 languages
- somewhat curated training rather than a raw flood of internet text (unclear if this curation helps or hinders performance, but it is there)
- emphasis on "IQ 100" level training not advanced training or reasoning. Apple wants the model to answer sensibly if you want to split tips, but does not [for now...] care how it responds if you give it your calculus homework

3. BY FAR the most important difference of these models, compared, to other models, is that they have been trained to handle a very specific task: a developer can, to simplify immensely, using natural and simple Swift, construct a query to be given to the LLM, and get back a response that is defined in terms of the structs and APIs of the calling app. This is not the same thing as "ask the LLM a question and get back text", and it's not the same thing as " ask the LLM a question and it gives you code that, hopefully, you can compile to do what you want".
The idea is that in a random app, like Uber Eats, I can say "I'd like to order that food we had two weeks ago, it was Asia, Thai I think, but I can't remember the name" and this will result in Uber Eats throwing up a useful order that you can click on. Look at what's involved here: the query has to go into the LLM (so that it can be "understood", Uber Eats also has to provide the database of recent orders (so that the LLM has a clue what was ordered over the relevant time period) and the response can't just be a text string like "looks like you ordered Satay Chicken, and by the way that's Indonesian not Thai", it has to be some sort of structure that plugs into Uber Eats API's to allow the construction of a genuine order.

No-one else has anything like this. THIS is what I mean when I say that Apple is constructing an AI UI, not an AI model.

4. So the position Apple finds itself in is that
- it's trying to figure out how to utilize LLMs AS API.
- this is a research task, so it makes no sense to try to do this research at the same time as you're constantly modifying an LLM that takes $100M per training run! Instead you fiddle around with tiny models, until you think you have the basic concepts required for the APIs (and their pieces within the OS runtime and Swift) all working.
- then you scale this up to a mid-sized model and validate that it still works.
- are they scaling it up to a much larger model? Maybe? Or maybe there is no point in doing that until we get a year or so of experience with this machinery to see what needs to be improved, changed, or discarded?


Apple is not playing the same game as Gemini, OpenAI etc. And it doesn't need to, just like Apple is not trying to compete with Google as a search engine. As long as the leading edge LLMs continue to provide as good an experience on Apple as they do anywhere else, then no-one feels any need to stop buying Apple HW just to get "optimal" ChatGPT.

This is all described in

Aspects of the Swift API were demo'd and examples given at multiple talks at WWDC earlier this year.
Apple did something very stupid last year with the announcement of Apple Intelligence before it was ready and before all this infrastructure was in place. That tells us that Apple Marketing were stupid in this case, and should have followed their normal rule of **** until you're ready to ship. But Apple Marketing is not Apple Engineering, and Apple Engineering have a plan in place grander than anything you can imagine.
Excellent post!

But I have to disagree on one point. Apple isn’t playing the same game as Gemini or OpenAI, not because they don’t want to, but because the competition has already pulled ahead. Just look at how well Gemini integrates with Android and Google Workspace, or how OpenAI / Copilot integrate seamlessly with Microsoft 365, Azure, and Dynamics. Now compare that to what we currently have with Apple Intelligence and the ChatGPT integration on iOS, iPadOS, and macOS.

This is the same company that acquired Siri 15 years ago, yet they never really pushed it forward in a meaningful way.

Your post highlights a lot of great ideas for how Apple could improve in the AI space, but the reality is the competition isn’t slowing down. The next few years are going to be very interesting.
 
Ollama models are objectively better. The only reason why I can see Apple using Gemini is so they can make money off the users' data somehow. So much for "If you don't pay for the precut, you are the product". This should also apply to search engines and AI. At least Meta is transparent on how they make money: By selling ads.
 
It’s only been a week and here we are getting the announcement of another red carpet being rolled out … this time for a privacy invader by a privacy protector who wants to win the AI Nobel Price at all cost.
 
Here is a screenshot I took a little while ago. I’m not sure whether to laugh cry!
 

Attachments

  • IMG_4451.jpeg
    IMG_4451.jpeg
    567.2 KB · Views: 19
To be more correct, they weren’t talking about Gemini, which is the cloud-based AI service operated by Google. Per the article, they were talking about training (more likely fine tuning) a *model* - which would be a Gemma model, which is what Gemini is based on.

These are not the same thing, and have wildly different implications. Apple has done work with open weight models such as Gemma, Mistral and others in the past-most notably, the open source coding model that they *taught* Swift programming to. I wouldn’t be surprised if they are doing something similar with Gemma.

Especially with the court cases around Google’s search placement fee. This would be a good way for Google to pay Apple without involving actual money, or affecting Apple’s privacy-centric approach at all.
 
It’s only been a week and here we are getting the announcement of another red carpet being rolled out … this time for a privacy invader by a privacy protector who wants to win the AI Nobel Price at all cost.
The conflation of Google training a model and Gemini somehow being involved is erroneous. Google’s models are called Gemma, not Gemini. Different things -and only a fine tuned Gemma makes sense here.
 
Why would Apple want to integrate a manipulated AI model that heavily leans into conspiracy theories and neonazism? Apple should indeed stay as far away as possible.
I’m sure Gemini has had no problems of its own. How are those glued pizzas with small rocks?
 
Apple Intelligence will never be able to stand on its own, they neglected research for too long and now they have accumulated a technological gap that will never be filled, especially considering the few employees having some knowledge in the topic are being poached by other companies. That being said grok and musk are toxic at this point, Apple should stay clear.
Too much R&D dollars spent on refining the poop emoji. True brilliance the way that emoji turned out!
 
Not that insane. Apple has a long tradition of partnering with other companies. For a long time , Apple relied on Google to develop their iOS map app. Eventually , they ended up making their own map app, but it took them almost a decade.

Before Apple made iCloud, Steve Jobs wanted to buy Dropbox but they refused his offer.
Apple didn’t have cellular modem expertise so they bought the Intel Modem division for $1 billion and hired all of Qualcomm’s top cellular modem engineers. Apple will just buy you up 💰
 
Just a funny moment... I looked at my Siri Suggestions before moving over to Samsung. One of the recommendations was to "Remember to run the bath faucet." What in the world? Maybe Siri can smell. I will say I use Gemini multiple times every day and it has been fast and fantastic. Apple has A LOT of catching up to do.
 
How does a company sitting on a trillion dollars of cash fall so far behind in the AI race? Genuine question. Did everything go into that absolute FLOP of a product known as Vision Pro?
Apple had there focus mainly on Apple Vision Pro and the Apple Car Project. Cook even admitted that the Vision Pro was a Niche product and would not sell a ton in the beginning. They had top secret labs 🥼 on the Vision Pro and should’ve had top secret labs 🧪 🧫 on AI artificial intelligence.
 
Please don't talk about things you know nothing about.

1. Why is Apple working with Google in this space? Because Apple trains its (MULTIPLE) models on TPUs.

2. Apple has at least two known "sexy" models, along with multiple more specialized models. The two sexy models are called AFM (Apple Foundation Model). They are both *combined* language and vision models.
The on-device model is ~3B parameters, quantized to 2b per parameter.
The server model runs on Apple HW (the GPU of what are presumably M2 or M3 Ultras) and has about 200B models, quantized to about 3.5b per parameter.
Both are competitive (not obviously worse or obviously better) than counterparts *at this size* (which is definitely not the largest size, guesses are eg chatGPT5 is ~8 times larger).

Unusual (or at least non-obvious) aspects of these models include
- combined text and vision rather than two separate models
- multilingual, handling about 15 languages
- somewhat curated training rather than a raw flood of internet text (unclear if this curation helps or hinders performance, but it is there)
- emphasis on "IQ 100" level training not advanced training or reasoning. Apple wants the model to answer sensibly if you want to split tips, but does not [for now...] care how it responds if you give it your calculus homework

3. BY FAR the most important difference of these models, compared, to other models, is that they have been trained to handle a very specific task: a developer can, to simplify immensely, using natural and simple Swift, construct a query to be given to the LLM, and get back a response that is defined in terms of the structs and APIs of the calling app. This is not the same thing as "ask the LLM a question and get back text", and it's not the same thing as " ask the LLM a question and it gives you code that, hopefully, you can compile to do what you want".
The idea is that in a random app, like Uber Eats, I can say "I'd like to order that food we had two weeks ago, it was Asia, Thai I think, but I can't remember the name" and this will result in Uber Eats throwing up a useful order that you can click on. Look at what's involved here: the query has to go into the LLM (so that it can be "understood", Uber Eats also has to provide the database of recent orders (so that the LLM has a clue what was ordered over the relevant time period) and the response can't just be a text string like "looks like you ordered Satay Chicken, and by the way that's Indonesian not Thai", it has to be some sort of structure that plugs into Uber Eats API's to allow the construction of a genuine order.

No-one else has anything like this. THIS is what I mean when I say that Apple is constructing an AI UI, not an AI model.

4. So the position Apple finds itself in is that
- it's trying to figure out how to utilize LLMs AS API.
- this is a research task, so it makes no sense to try to do this research at the same time as you're constantly modifying an LLM that takes $100M per training run! Instead you fiddle around with tiny models, until you think you have the basic concepts required for the APIs (and their pieces within the OS runtime and Swift) all working.
- then you scale this up to a mid-sized model and validate that it still works.
- are they scaling it up to a much larger model? Maybe? Or maybe there is no point in doing that until we get a year or so of experience with this machinery to see what needs to be improved, changed, or discarded?


Apple is not playing the same game as Gemini, OpenAI etc. And it doesn't need to, just like Apple is not trying to compete with Google as a search engine. As long as the leading edge LLMs continue to provide as good an experience on Apple as they do anywhere else, then no-one feels any need to stop buying Apple HW just to get "optimal" ChatGPT.

This is all described in

Aspects of the Swift API were demo'd and examples given at multiple talks at WWDC earlier this year.
Apple did something very stupid last year with the announcement of Apple Intelligence before it was ready and before all this infrastructure was in place. That tells us that Apple Marketing were stupid in this case, and should have followed their normal rule of **** until you're ready to ship. But Apple Marketing is not Apple Engineering, and Apple Engineering have a plan in place grander than anything you can imagine.
You're absolutely right. It's pretty crazy how many people forget that Apple has been at the forefront of Machine Learning for a very long time. I haven't checked it recently, but awhile back, Apple put a lot of early small models on Hugging Face that people could download and try themselves. They're not ignorant of the AI or model training space at all, they've been pioneers in it. They just haven't gotten to a point where they are comfortable shipping what they have, and so are leaning on these third parties in the meantime.

Does Tim Cook really hate Elon so much that’s he’d rather partner with the tech devil than temporarily lean on Grok until Apple Intelligence can stand on its own? 🤔
It definitely shouldn't be ruled out, it's arguably the most advanced model publicly available today and like the others is continually improving. However, it's not really an "Apple" type model. Apple is all about presenting a "safe" and "family friendly" face to things, and a model that's designed to be somewhat less censored than most others (although still heavily censored) like Grok isn't going to be at the top of their list. Apple's not going to be interested in shipping a Siri-like replacement product without extreme censorship, and xAI may not be willing to devote resources to training a special and "safely" censored model just for Apple.
 
I wouldn't be surprised if some government in the future forced apple and other companies to do just that, if apple doesn't do it on their own.
I wouldn’t be shocked if Apple merges with google in the next ten years, with google as the senior partner in this.

Apple for the hardware & google for the cloud, software - and AI.
 
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.