Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.
What Apple inspire in a lot of everything they do, they surely take back through Siri.
I turned her off a long time ago, but I can still feel the irritation of all the endless errors.
It's a disaster that a company with Apple's magnitude continues doing something so useless.
 
Of course but I’m not talking about degree or two Fahrenheit. It’s off by A LOT! Last night it said it was 10° F warmer here than any other weather app I use, which were all within 2 degrees of each other. I’m not entirely hating on it, I love the layout, the types of info provided and some other features… I just want the info correct.
How do you know those other apps were correct? My back porch thermometer currently says 27 and Apple's weather app says 30. But Apple's weather app is just giving a general number for the city I live in which is rather large geographically and has a lot of elevation changes. It's not a big deal.
 
"Hey Siri, who won the Super Bowl in 1996?"...Siri says the Dallas Cowboys which is correct.
Exactly. It's mostly a wording issue. Granted the rudimentary LLM should interpret Super Bowl 60 and the 60th Super Bowl the same (or "Super Bowl in 1996", "Super Bowl XXX", "Super Bowl 30", or "30th Super Bowl"), but wording does matter.

The same thing happens with people as well. If you word questions differently on surveys or questionnaires, you can get drastically different answers. This is well-established for people: https://www.pewresearch.org/writing-survey-questions/

Siri(GPT?) is inconsistent in a way that many people would not be with a relatively trivial difference between Super Bowl 30 or 30th Super Bowl; however, that it is inconsistent based on how you word the questions actually makes Siri's responses more human-like, which is kind of funny.
 
Tim Cook has overall done a fine job, but his time is over. It's time for a new CEO with fresh ideas and a spirit of innovation and demand for quality that seems lost.

I agree with this, totally.

It has been a phenomenal run and he did a fantastic transition from Steve to not have any disruption and really take things over the moon with the trajectory that had begun.

It’s been a financial windfall…. But we should also keep in mind the entire tech market had an unbelievable decade long bull run on the whole. It has not been just Apple or anything.

The concerning thing is the track record on product, which really speaks for itself the last five years plus

If they don’t make a change soon, it is going to manifest itself in a disaster a few years down the line
 
“Siri repeatedly and incorrectly credited the Philadelphia Eagles with 33 Super Bowl victories, despite the team having won only one championship in their history.”

The Eagles have four NFL championships. Three of them are prior to the Super Bowl.
 
Maybe Apple use Siri for subliminals to influence people to buy more Apple stuff 😉
 
How do you know those other apps were correct? My back porch thermometer currently says 27 and Apple's weather app says 30. But Apple's weather app is just giving a general number for the city I live in which is rather large geographically and has a lot of elevation changes. It's not a big deal.
Because they agreed with my porch thermometer
 
This is interesting but would be much more interesting asking Android same questions, asking ChatGPT same questions. ChatGPT lies to me all the time about facts. Its crazy.
 


In what may not come as much of a surprise, a new test of Siri's knowledge of Super Bowl history has revealed significant accuracy issues with Apple's virtual assistant, suggesting Apple still has some way to go in overcoming challenges with Siri's ability to provide reliable information.

Should-Apple-Kill-Siri-Feature.jpg

In a methodical experiment, One Foot Tsunami's Paul Kafasis asked Siri who won each Super Bowl from I through LX and documented its responses. The results were strikingly poor, with Siri correctly identifying winners only 34% of the time – just 20 correct answers out of 58 played Super Bowls.

Perhaps most notably, Siri repeatedly and incorrectly credited the Philadelphia Eagles with 33 Super Bowl victories, despite the team having won only one championship in their history. The virtual assistant's responses ranged from providing information about wrong Super Bowls to offering completely unrelated football facts.

While Siri did manage a few streaks of accurate answers, including three consecutive correct responses for Super Bowls V through VII, it also had a remarkable string of 15 consecutive incorrect answers spanning Super Bowls XVII through XXXII.

In one telling instance, when asked about Super Bowl XVI, Siri offered to defer to ChatGPT - which then provided the correct answer. The contrast highlighted the limitations of Siri's own knowledge base compared to more advanced AI systems.

The test was conducted on iOS 18.2.1 with Apple Intelligence enabled, and similar results were found on both the upcoming iOS 18.3 beta and macOS 14.7.2, suggesting the issue extends across Apple's platforms. Kafasis generated a spreadsheet of the results in both Excel and PDF formats, which you can read here.

Separately, inspired by Kafasis' test, Daring Fireball's John Gruber tried some of his own sports queries with Siri and compared its responses to ChatGPT, Kagi, DuckDuckGo, and Google, all of which succeeded where Siri failed.

Perhaps worse for Apple, Gruber found that old Siri (i.e. before Apple Intelligence) did a better job at answering a question by declining to answer it, instead providing a list of web links. The first web result provided an accurate, if only partial, answer to the question, whereas new Siri, powered by Apple Intelligence, fared much worse. Gruber explains:
"It's just incredible how stupid Siri is about a subject matter of such popularity," commented Gruber. "If you had guessed that Siri could get half the Super Bowls right, you lost, and it wasn't even that close."

Of course, this isn't the first time Siri has received heavy flak for its all-round performance, but Gruber's criticism about "plausibly wrong" answers to general knowledge questions ties back to the modern problem of hallucinating AI chatbots that spout misleading or flat-out wrong responses with complete confidence.

Apple is developing a much smarter version of Siri that utilizes advanced large language models, which should allow the personal assistant to better compete with chatbots like ChatGPT. A chatbot version of Siri would likely be able to hold ongoing conversations and provide the sort of help and insight as ChatGPT or Claude, but how well the integration will perform may be a concern, going on Siri's abysmal track record.

Apple is expected to announce LLM Siri as soon as 2025 at WWDC, but Apple won't launch it until several months after it's unveiled. That means LLM Siri would come in an update to iOS 19, with Apple planning for a spring 2026 launch.

Article Link: Siri Gives Eagles 33 False Super Bowl Wins in Basic Knowledge Test
Unable to reproduce this result. Siri gets it right every time for me on my iPhone 16 Pro Max on iOS 18.3 RC.

Of course, I use Siri a lot, for fairly long conversations where I clarify if she answered wrong, assuming she misunderstood me, same as if she was a human assistant.
 

Attachments

  • IMG_1425.jpeg
    IMG_1425.jpeg
    205 KB · Views: 24
  • IMG_1424.jpeg
    IMG_1424.jpeg
    209.3 KB · Views: 28
  • IMG_1426.jpeg
    IMG_1426.jpeg
    206.1 KB · Views: 23
  • IMG_1423.jpeg
    IMG_1423.jpeg
    224.2 KB · Views: 25
That's pretty bad. I use a LLM locally (hosted on a headless PC in my office), and it did great on the Super Bowl questions (I only asked 8, spread around). Pretty sad that my little setup is doing better than Siri/ChatGPT.
 
Anyone who uses Siri for anything other than setting a timer is either brave or stupid 😂
I don’t think I am either one of those, but Siri is very helpful at turning my lights off and on, which is difficult for me due to nerve damage in my hand and fingers. She’s also pretty decent about the reminders that I have programmed into her Shortcuts to make me remember to take my medicine and check on the temperature in the winter so that if it’s too cold, I can take precautions so my pipes don’t freeze. Really, because I’m retired, she competently does a lot of the things that I need her for, given that my needs aren’t demanding, but I’m ready for her to be more like Jarvis, though not sentient.
 
  • Like
Reactions: robvalentine
iOS is stable, but I would not say it is better than barebones Android (no skin) feature-wise.

MacOS is also stable, but not necessarily better than Windows in many ways. There are things that are easier in Windows, and for fields like Civil Engineering, Mechanical Engineering, Computational Fluid Dynamics Windows is superior, more compatible with workflow and most software won’t run on MacOS unless you virtualize it and it becomes a slug fest. Gaming is also better on Windows.
We are talking about Operating Systems, not about what third party software runs on it.
Windows is still far behind and too many unsolved bugs in it.
 
Apple is a Hardware Company. Always has been. They give away the software so you buy the expensive hardware. They have no core competency for software. And AI is all about software.
 
I don’t think I am either one of those, but Siri is very helpful at turning my lights off and on, which is difficult for me due to nerve damage in my hand and fingers. She’s also pretty decent about the reminders that I have programmed into her Shortcuts to make me remember to take my medicine and check on the temperature in the winter so that if it’s too cold, I can take precautions so my pipes don’t freeze. Really, because I’m retired, she competently does a lot of the things that I need her for, given that my needs aren’t demanding, but I’m ready for her to be more like Jarvis, though not sentient.
I agree with you, Siri is generally, pretty good for basic tasks like these. 99/100 when I run an automation to turn off the lights and lock the doors before I go to bed, it works great.

It's that one time when I give it the same old command, and it gives me some nonsense answer. "Good night," I say. "Night night." Ugh.

It's the inconsistency that kills me, and there's no way to debug it.

I haven't adapted yet to using Siri for general knowledge questions. I still use the web for that, which is getting less and less useful as AI results dominate the top results.

I'm curious to see how the 18.4 betas improve this experience. It could be terrible to start, maybe even worse than the status quo, because it's a beta.
 
Also...it makes a big difference if these were typed vs voice. Voice is always a problem for every system not just siri.

I just typed to Siri and it got the majority of ones I checked fine until it thought I was asking about recent scores then it just gave me most recent score over and over.

ChatGPT got everything I typed fine.
 
Is it Philadelphia day? I saw that 76% of iPhones from the last four years are running iOS 18 and then, the Eagles won 33% of Super Bowls. Maybe, it's just that Siri is a Philadelphia fan.

I tried Siri vs Google Assistant in 2018 and Siri could barely handle anything while Google Assistant got everything right. It's unfortunate that it isn't better.
 
Sounds like it would be easier to just google it yourself 🤷🏻‍♂️
 
Not surprised. Siri has never been intelligent. Hopefully that changes soon. Apple may announce something at WWDC but a lot of work is to be done on Siri.
 
  • Like
Reactions: mganu
No AI is intelligent and does not think, although the impression is completely different. No LLM is suitable for searching, although GPT is the closest. All LLMs are based on probability. If you ask a question, there is no reasoning in it, and the result is that next to the word X, Y occurs most often. The more words, the more variables, but also the greater the probability of a good answer. However, no LLM understands the question and does not understand the answer. That is why there are so many hallucinations, because the LLM does not know that it does not know.
 
So the false Philadelphia Eagles results are Apple's pre-approved, biased and politically correct version of the truth? That is silly!
You know some brain power would be helpful here.

If Apple can get these types of basic provable facts wrong, think how misguided their gray area summaries, opinions, and political reporting is.
 
You know some brain power would be helpful here.

If Apple can get these types of basic provable facts wrong, think how misguided their gray area summaries, opinions, and political reporting is.
Insulting me doesn't make you right, it makes you impolite.

Making mistakes is not being pre-approved, biased and politically correct.
 
I don’t think I am either one of those, but Siri is very helpful at turning my lights off and on, which is difficult for me due to nerve damage in my hand and fingers. She’s also pretty decent about the reminders that I have programmed into her Shortcuts to make me remember to take my medicine and check on the temperature in the winter so that if it’s too cold, I can take precautions so my pipes don’t freeze. Really, because I’m retired, she competently does a lot of the things that I need her for, given that my needs aren’t demanding, but I’m ready for her to be more like Jarvis, though not sentient.
Fair enough. I just find she struggles to switch my front door light on and off. She says I need to download an app for it. I just navigate to the home app and do it that way.
 
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.