Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.
Imagine holding a summit for something you barely do, much less do well.

A lot of commenters on here seem to think AI is just about ChatGPT/GPT-4/DALL-E/Stable Diffusion/MidJourney... which isn't true at all. Apple has already been doing a lot of AI and doing it well. It's just not obvious because it's baked into their products. And I'm not talking about Siri.

  • FaceID is powered by AI. Last I checked that do that extremely well.

  • Handwriting recognition on the iPad when writing with the Apple Pencil is powered by AI. It does an amazing job.

  • The voice-powered translation in the Translate app uses on-device ML and AI to do real-time translations. It works great.

  • The camera on the iPhone uses AI and ML with every photo you take in order to produce better results. Portrait mode, cinema mode, etc. all leverage AI and ML. And do it on device and incredibly fast. Last I checked those were excellent.

  • The object and people recognition in the Photos app for searching your photo library is fantastic. Again, powered by AI and ML.

AI and ML capabilities are baked into Apple's Silicon hardware. And again... we know Apple Silicon is excellent.

Apple is already doing AI. And doing it well. So to pretend otherwise is disingenuous.
 
I really hope product launches stick to the recorded format… they’re so much slicker and get through things in a much faster way
 
Early on I found siri very unhelpful so I never use it.

I find it useful for a few very specific things. Mostly "set a timer for X minutes", "remind me to...", and "set my alarm for ...". These tend to be done faster with Siri than by opening the specific apps and poking away. Can even do them from my Apple Watch if I don't have the phone to hand.

If Siri supported Spotify then that would be super useful too. Google/Alexa are good for "play X on Spotify" on smart speakers, etc. It would be great to be able to do that on Apple devices.
 
  • Like
Reactions: surfzen21
Handwriting recognition on the iPad when writing with the Apple Pencil is powered by AI. It does an amazing job.
The voice-powered translation in the Translate app uses on-device ML and AI to do real-time translations. It works great.
The object and people recognition in the Photos app for searching your photo library is fantastic. Again, powered by AI and ML.
Agree to disagree.
 
Agreed! Going to miss these Apple transitions. They were fun!

View attachment 2154845

View attachment 2154848


But then you miss out on the groans when Apple announces a $1000 monitor stand, etc.

Sometimes the live audience reactions are the most honest, un-ignorable feedback a company ever gets.

"Do you guys not have phones?"

And you can bet they'll still be cutting to mini-movies of Hair Force One. Now that we don't have Jony's spaceship anymore they have to have something.
 
Neither of them are ‘smart’.

They are programs running on language models.

It is the people who develop them who are ‘smart’.

Actual people.

As long as we remember that.

This seems to be the key thing people keep forgetting. These AIs have to be constantly fed new data. That data has to come from people. If it's AIs feeding AIs it's just going to result in worse and worse copies.
 
Interesting how this chatGPT now sounds, as it has been created by Μ$… It wasn't, they just buy, they never did a good product… The last time I checked openAI chatGPT got some investors, named, Microsoft, NVIDIA, Amazon, Elon Musk, and Apple. Now, who knows the future, who would say in some years we will be laughing off for chatGPT to be so poor…
 
  • Love
Reactions: wilhoitm
This seems to be the key thing people keep forgetting. These AIs have to be constantly fed new data. That data has to come from people. If it's AIs feeding AIs it's just going to result in worse and worse copies.

Everything starts from people because we build them, code them, curate them, bug fix them, maintain them, and we are the ultimate user and consumer.

People who are enamoured and brainwashed by “AI“ are no different from people who believe David Blaine really does magic. They think it’s some kind of magical ghost inside a box. It’s just a bunch of statements and conditions hooked into databases (ok I know there is a little more :p)

Once the hype passes and this stuff becomes normal nobody will give AF. It will just be like using long form auto type or procedural graphics generators.

We will still be the users figuring out the front ends, frustrated with the outputs, facepalming at the bugs, complaining to developers about crashes and freezes, making jokes and memes about expectation vs reality.

I won’t be affected at all, most people won’t. We will still prefer to listen to people, read real authors, watch movies by really great artists. That’s what fandom is about. It’s about respecting people who struggled to make something.
 
  • Like
Reactions: Arran
Siri is just a frontend. The search itself uses Bing, and Microsoft is going to marry Bing with ChatGPT, so Siri will get some benefit as well.

Apple outsourcing even their "AI"... 🤣
I've never had many issues with searches, except she almost always sends stuff to my phone. If I wanted to use my Phone, I would have. The issue is her inconsistency with controlling HK Accessories. But now Google started doing all kinds of BS too, so I switched back to the HP Minis and lights, and it seems OK for now. She is even snappier about doing the things now too. The one true advantage with HP Mini vs Google Mini is her mic. Even though it can get annoying for the LR one to answer when I'm literally in the BR talking normal, at least I don't have to yell for her to respond, and she understands better. And she seems to be fixed about recognizing our AM sub too. But how many days now until she deteriorates again LOL
 
Siri is just a frontend. The search itself uses Bing, and Microsoft is going to marry Bing with ChatGPT, so Siri will get some benefit as well.

Apple outsourcing even their "AI"... 🤣

This (and most of these) comments are extremely silly.
An LLM (large language model) is essentially a language "extrapolator" -- given this string of text, extrapolate some successor text that is statistically plausible. Without wanting to get into issues of "what is intelligence", no-one in the know is even claiming this, *by itself*, counts as intelligence.
This sort of machinery is useful for many things (it does an adequate job of translation, or "create text in the style of"), but it is not designed to answer questions, and it's not even optimal for some of the target tasks like "summarize this text".

Consider now something like Siri (or a generic "assistant"). What it needs is
- an audio model to understand the noise it is hearing and convert it to text.
This is the first (and I suspect frequently the most common) place where things fall apart. VERY good, multiple, mics are required to get an acceptable audio signal, and such mics are not present in many situations. CarPlay (using the car mic) is one such, old macs (only one mic) is another. Having an AirPod in your ear is a 3rd case - the AirPod mic for a single AirPod seems to be lousy, which isn't helped by Apple seeming to have a bug where it can't decide whether to use the AirPod, aWatch, or iPhone mic.

- an LLM to convert the text to some sort of "understanding" of what needs to be done. IMHO Apple actually does this part reasonably well.

- an implementation of the "understanding" from the previous step. This part Apple does very badly. Some of the bad implementation is pure stupidity; things like saying "I can't do X on a Watch [or HomePod]" rather than just fscking sending the request to my iPhone or Mac, and I think this, such obvious, stupidity, is the largest single driver of Siri anger.
Another part of the bad implementation is requests for some sort of data lookup. This requires a combination of a web lookup and then perhaps some massaging of the results. This part Apple also does poorly, but I can't tell if it's the web lookup or the massaging that is the larger problem.

- finally what's needed is some sort of on-going history. Both short term memory (understand how my current sentence relates the previous sentence) and long term memory (behave like a human assistant so that you know if I usually refer to Debby in contexts about my personal life, I am referring to my wife, and you don't need to ask me, every time, if I might mean some other Debby in my address book that I last interacted with ten years ago!)

There's a lot of moving pieces here, and simply saying "ChatGPT awesome" doesn't solve most of them!
If I were Apple, apart from the "AI" (mainly people currently think LLM) parts I would work very hard on
- better audio models. There are multiple places where people still interact with bad mics, and it's worth putting more compute effort into processing that lousy signal

- better user feedback. There is far too much of things like Apple Watch for whatever reason can't talk to iPhone, or iPhone doesn't have the right internet connection or whatever, and you say your "Hey Siri, Remind me to xyz" only to be told "I can't do that right now".
Make a BETTER UI for these sorts of failure cases!!! Remember the audio data for what I told you and try again when you do have a connection? If the audio signal is too lousy to process locally, send it to some Apple computer to process. Worst case, just remember it so I can replay it as audio.
Another failure mode that people HATE HATE HATE is when you're in the car, you say "Hey Siri remind me to xyz" and Siri creates a Reminder to "kjhjkh57%%^%hjhjh", ie some garbled garbage that you have no idea what it was. There should be a way to replay the audio that generated the Reminder (or a Note) so you can figure out what you were trying to remind yourself to do.

- a better indication of what Siri thought you wanted it to do. If you look at how Wolfram Alpha handles this, you can make a natural language request, and Alpha gives its interpretation of what it thought you meant, along with multiple possible answers to different aspects of the question. Apple (and other assistants) are often too focussed on assuming there is one single response to an assistant and don't even consider giving two or four responses, and an explanation of what the assistant was trying to do.

- too much obsession with voice only. Voice has its place, but I want an Assistant not a Voice assistant! I want to be able to interact with a previous voice response using typing, or touch/pointer, or via cut and paste, or via dragging a file. Apple (and everyone else, but forget them, they are hopeless) has been locked too long in the model, which made sense in 2010 but not in 2023, of Siri as a voice statement/response system, not Siri as a high quality TOTAL assistant...
 
  • Like
Reactions: insomniac86
This (and most of these) comments are extremely silly.
An LLM (large language model) is essentially a language "extrapolator" -- given this string of text, extrapolate some successor text that is statistically plausible. Without wanting to get into issues of "what is intelligence", no-one in the know is even claiming this, *by itself*, counts as intelligence.
This sort of machinery is useful for many things (it does an adequate job of translation, or "create text in the style of"), but it is not designed to answer questions, and it's not even optimal for some of the target tasks like "summarize this text".

Consider now something like Siri (or a generic "assistant"). What it needs is
- an audio model to understand the noise it is hearing and convert it to text.
This is the first (and I suspect frequently the most common) place where things fall apart. VERY good, multiple, mics are required to get an acceptable audio signal, and such mics are not present in many situations. CarPlay (using the car mic) is one such, old macs (only one mic) is another. Having an AirPod in your ear is a 3rd case - the AirPod mic for a single AirPod seems to be lousy, which isn't helped by Apple seeming to have a bug where it can't decide whether to use the AirPod, aWatch, or iPhone mic.

- an LLM to convert the text to some sort of "understanding" of what needs to be done. IMHO Apple actually does this part reasonably well.

- an implementation of the "understanding" from the previous step. This part Apple does very badly. Some of the bad implementation is pure stupidity; things like saying "I can't do X on a Watch [or HomePod]" rather than just fscking sending the request to my iPhone or Mac, and I think this, such obvious, stupidity, is the largest single driver of Siri anger.
Another part of the bad implementation is requests for some sort of data lookup. This requires a combination of a web lookup and then perhaps some massaging of the results. This part Apple also does poorly, but I can't tell if it's the web lookup or the massaging that is the larger problem.

- finally what's needed is some sort of on-going history. Both short term memory (understand how my current sentence relates the previous sentence) and long term memory (behave like a human assistant so that you know if I usually refer to Debby in contexts about my personal life, I am referring to my wife, and you don't need to ask me, every time, if I might mean some other Debby in my address book that I last interacted with ten years ago!)

There's a lot of moving pieces here, and simply saying "ChatGPT awesome" doesn't solve most of them!
If I were Apple, apart from the "AI" (mainly people currently think LLM) parts I would work very hard on
- better audio models. There are multiple places where people still interact with bad mics, and it's worth putting more compute effort into processing that lousy signal

- better user feedback. There is far too much of things like Apple Watch for whatever reason can't talk to iPhone, or iPhone doesn't have the right internet connection or whatever, and you say your "Hey Siri, Remind me to xyz" only to be told "I can't do that right now".
Make a BETTER UI for these sorts of failure cases!!! Remember the audio data for what I told you and try again when you do have a connection? If the audio signal is too lousy to process locally, send it to some Apple computer to process. Worst case, just remember it so I can replay it as audio.
Another failure mode that people HATE HATE HATE is when you're in the car, you say "Hey Siri remind me to xyz" and Siri creates a Reminder to "kjhjkh57%%^%hjhjh", ie some garbled garbage that you have no idea what it was. There should be a way to replay the audio that generated the Reminder (or a Note) so you can figure out what you were trying to remind yourself to do.

- a better indication of what Siri thought you wanted it to do. If you look at how Wolfram Alpha handles this, you can make a natural language request, and Alpha gives its interpretation of what it thought you meant, along with multiple possible answers to different aspects of the question. Apple (and other assistants) are often too focussed on assuming there is one single response to an assistant and don't even consider giving two or four responses, and an explanation of what the assistant was trying to do.

- too much obsession with voice only. Voice has its place, but I want an Assistant not a Voice assistant! I want to be able to interact with a previous voice response using typing, or touch/pointer, or via cut and paste, or via dragging a file. Apple (and everyone else, but forget them, they are hopeless) has been locked too long in the model, which made sense in 2010 but not in 2023, of Siri as a voice statement/response system, not Siri as a high quality TOTAL assistant...

Let’s just remember all this comes at a large computational cost that eats into profits.

There could also be knock on effects that hurt society.

With all the power demands of LLMs and other “AI” models, the costs of wholesale energy could increase and that means energy costs rising for everyone.

We aren’t even close to renewable energy for all, and renewables are not cheap anyway.

So there is a large carbon footprint to consider and if we ask those people at OpenAI or Midjourney about their carbon footprint, e-waste and costs….they will remain silent because the answer isn’t pleasant.
 
  • Disagree
Reactions: name99
Imagine holding a summit for something you barely do, much less do well.
Apple is actually an industry leader in the implementation and deployment of AI. iOS and Mac OS use tons of AI and it’s deployed on hundreds of millions of devices globally. They just don’t make it a flashy talking point.
 
Last edited:
  • Like
Reactions: Arran and name99
A lot of commenters on here seem to think AI is just about ChatGPT/GPT-4/DALL-E/Stable Diffusion/MidJourney... which isn't true at all. Apple has already been doing a lot of AI and doing it well. It's just not obvious because it's baked into their products. And I'm not talking about Siri.

  • FaceID is powered by AI. Last I checked that do that extremely well.

  • Handwriting recognition on the iPad when writing with the Apple Pencil is powered by AI. It does an amazing job.

  • The voice-powered translation in the Translate app uses on-device ML and AI to do real-time translations. It works great.

  • The camera on the iPhone uses AI and ML with every photo you take in order to produce better results. Portrait mode, cinema mode, etc. all leverage AI and ML. And do it on device and incredibly fast. Last I checked those were excellent.

  • The object and people recognition in the Photos app for searching your photo library is fantastic. Again, powered by AI and ML.

AI and ML capabilities are baked into Apple's Silicon hardware. And again... we know Apple Silicon is excellent.

Apple is already doing AI. And doing it well. So to pretend otherwise is disingenuous.

Then how come I can't ask Siri what 2+2 is and get an answer without an internet connection?

Serious question. If they're so serious about AI where is the offline Siri they've been promising the last couple years?

Not criticising you, you're not wrong. I just don't know why they keep pretending Siri is just fine.
 
This has little to no bearing on the public events.

This is an internal event and most likely was forced in person to try to strengthen the return to office push they have tried and failed to make. I know my office is trying to persuade us to come into the office with ice cream gatherings and ‘meet X or Y’ leader.

On a side note, Apple better be planning more than a summit in this space as they are away to get spanked if they don’t.
 
This (and most of these) comments are extremely silly.
An LLM (large language model) is essentially a language "extrapolator" -- given this string of text, extrapolate some successor text that is statistically plausible. Without wanting to get into issues of "what is intelligence", no-one in the know is even claiming this, *by itself*, counts as intelligence.
This sort of machinery is useful for many things (it does an adequate job of translation, or "create text in the style of"), but it is not designed to answer questions, and it's not even optimal for some of the target tasks like "summarize this text".

Consider now something like Siri (or a generic "assistant"). What it needs is
- an audio model to understand the noise it is hearing and convert it to text.
This is the first (and I suspect frequently the most common) place where things fall apart. VERY good, multiple, mics are required to get an acceptable audio signal, and such mics are not present in many situations. CarPlay (using the car mic) is one such, old macs (only one mic) is another. Having an AirPod in your ear is a 3rd case - the AirPod mic for a single AirPod seems to be lousy, which isn't helped by Apple seeming to have a bug where it can't decide whether to use the AirPod, aWatch, or iPhone mic.

- an LLM to convert the text to some sort of "understanding" of what needs to be done. IMHO Apple actually does this part reasonably well.

- an implementation of the "understanding" from the previous step. This part Apple does very badly. Some of the bad implementation is pure stupidity; things like saying "I can't do X on a Watch [or HomePod]" rather than just fscking sending the request to my iPhone or Mac, and I think this, such obvious, stupidity, is the largest single driver of Siri anger.
Another part of the bad implementation is requests for some sort of data lookup. This requires a combination of a web lookup and then perhaps some massaging of the results. This part Apple also does poorly, but I can't tell if it's the web lookup or the massaging that is the larger problem.

- finally what's needed is some sort of on-going history. Both short term memory (understand how my current sentence relates the previous sentence) and long term memory (behave like a human assistant so that you know if I usually refer to Debby in contexts about my personal life, I am referring to my wife, and you don't need to ask me, every time, if I might mean some other Debby in my address book that I last interacted with ten years ago!)

There's a lot of moving pieces here, and simply saying "ChatGPT awesome" doesn't solve most of them!
If I were Apple, apart from the "AI" (mainly people currently think LLM) parts I would work very hard on
- better audio models. There are multiple places where people still interact with bad mics, and it's worth putting more compute effort into processing that lousy signal

- better user feedback. There is far too much of things like Apple Watch for whatever reason can't talk to iPhone, or iPhone doesn't have the right internet connection or whatever, and you say your "Hey Siri, Remind me to xyz" only to be told "I can't do that right now".
Make a BETTER UI for these sorts of failure cases!!! Remember the audio data for what I told you and try again when you do have a connection? If the audio signal is too lousy to process locally, send it to some Apple computer to process. Worst case, just remember it so I can replay it as audio.
Another failure mode that people HATE HATE HATE is when you're in the car, you say "Hey Siri remind me to xyz" and Siri creates a Reminder to "kjhjkh57%%^%hjhjh", ie some garbled garbage that you have no idea what it was. There should be a way to replay the audio that generated the Reminder (or a Note) so you can figure out what you were trying to remind yourself to do.

- a better indication of what Siri thought you wanted it to do. If you look at how Wolfram Alpha handles this, you can make a natural language request, and Alpha gives its interpretation of what it thought you meant, along with multiple possible answers to different aspects of the question. Apple (and other assistants) are often too focussed on assuming there is one single response to an assistant and don't even consider giving two or four responses, and an explanation of what the assistant was trying to do.

- too much obsession with voice only. Voice has its place, but I want an Assistant not a Voice assistant! I want to be able to interact with a previous voice response using typing, or touch/pointer, or via cut and paste, or via dragging a file. Apple (and everyone else, but forget them, they are hopeless) has been locked too long in the model, which made sense in 2010 but not in 2023, of Siri as a voice statement/response system, not Siri as a high quality TOTAL assistant...
So basically Siri sucks
 
  • Disagree
Reactions: name99
The most recent official CDC weekly death toll? 3,452. Higher than 9/11. But back to "normal" for most people. The new normal being acceptance of never-ending mass death when we could simply mask, ventilate and vaccinate.

 
The most recent official CDC weekly death toll? 3,452. Higher than 9/11. But back to "normal" for most people. The new normal being acceptance of never-ending mass death when we could simply mask, ventilate and vaccinate.

They, uh, never offered more vaxxes chief. Blame the companies or the government, not “most people.”
 
Then how come I can't ask Siri what 2+2 is and get an answer without an internet connection?

Serious question. If they're so serious about AI where is the offline Siri they've been promising the last couple years?

Not criticising you, you're not wrong. I just don't know why they keep pretending Siri is just fine.
Because
- no-one has an *LLM* that reliably does arithmetic. Wolfram Alpha does, but it’s a different sort of creature…
- LLMs are LARGE. The AI models that Apple deploys on-machine are substantially smaller and substantially more limited;the largest are probably the image+text models that classify and label images, and these are a few 10s of MB.

This stuff is not secret if you switch off the snark long enough to spend a few hours reading the pages on the Apple Machine Learning website…
 
  • Like
Reactions: StyxMaker
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.