Oh yeah. There were some cases of that.Heck, in some areas it's actually third world cheap labor in a data center.
Oh yeah. There were some cases of that.Heck, in some areas it's actually third world cheap labor in a data center.
It's like ML with plugins. Actual reasoning/questioning depends on consciousness. I'm not sure it's possible on current hardware. I think it's somewhere in quantum future.Most current AI is actually a very sophisticated search engine capable of of combining information from many sources into one (largely) coherent answer. Still very useful, but important to understand the limitations.
Not true. However, it's true that they're not as intelligent as us and they clearly lack some architectural things they'll need. But the simple statements from people in this thread that LLMs have no intelligence is wrong.Nobody who understands LLMs actually thinks that they're intelligent. ..
It suggests that LLMs don't scale reasoning like humans do, overthinking easy problems and thinking less for harder ones.
When will it burst all I see is an obsession with ai in every industryOf course. “AI” is just a marketing term at this point, and not any kind of actual intelligence. These AIs are really just glorified search engines that steal peoples’ hard work and regurgitate that work as if the data is it’s own. We’re just living in an “AI bubble” that will burst sooner rather than later.
Nobody who understands LLMs actually thinks that they're intelligent … Apple is beginning to look like Microsoft and BlackBerry denying the potential of the iPhone.
Is anyone surprised Apple is downplaying AI, when their own Apple Intelligence is in the doldrums?
A newly published Apple Machine Learning Research study has challenged the prevailing narrative around AI "reasoning" large-language models like OpenAI's o1 and Claude's thinking variants, revealing fundamental limitations that suggest these systems aren't truly reasoning at all.
![]()
For the study, rather than using standard math benchmarks that are prone to data contamination, Apple researchers designed controllable puzzle environments including Tower of Hanoi and River Crossing. This allowed a precise analysis of both the final answers and the internal reasoning traces across varying complexity levels, according to the researchers.
The results are striking, to say the least. All tested reasoning models – including o3-mini, DeepSeek-R1, and Claude 3.7 Sonnet – experienced complete accuracy collapse beyond certain complexity thresholds, and dropped to zero success rates despite having adequate computational resources. Counterintuitively, the models actually reduce their thinking effort as problems become more complex, suggesting fundamental scaling limitations rather than resource constraints.
Perhaps most damning, even when researchers provided complete solution algorithms, the models still failed at the same complexity points. Researchers say this indicates the limitation isn't in problem-solving strategy, but in basic logical step execution.
Models also showed puzzling inconsistencies – succeeding on problems requiring 100+ moves while failing on simpler puzzles needing only 11 moves.
The research highlights three distinct performance regimes: standard models surprisingly outperform reasoning models at low complexity, reasoning models show advantages at medium complexity, and both approaches fail completely at high complexity. The researchers' analysis of reasoning traces showed inefficient "overthinking" patterns, where models found correct solutions early but wasted computational budget exploring incorrect alternatives.
The take-home of Apple's findings is that current "reasoning" models rely on sophisticated pattern matching rather than genuine reasoning capabilities. It suggests that LLMs don't scale reasoning like humans do, overthinking easy problems and thinking less for harder ones.
The timing of the publication is notable, having emerged just days before WWDC 2025, where Apple is expected to limit its focus on AI in favor of new software designs and features, according to Bloomberg.
Article Link: Apple Research Questions AI Reasoning Models Just Days Before WWDC
Which is strictly Cook not having any clue about Tech. He is a glorified accountant and needs to go.It’s pretty clear at this point that Apple execs bought into the AI hype until realizing that it was more hype than helpful. They just got ahead of themselves at last years’ WWDC before testing that the technology they were trying to incorporate into devices actually worked as they expected it to.
Which is strictly Cook not having any clue about Tech. He is a glorified accountant and needs to go.
The best part of Apple Intelligence has been the default config RAM spec bumps
I strongly recommend you and anyone reading this who cares watch Yann’s recent presentations from 2024 and 2025. All of them are worthwhile, and he’s one of the very few people not hyping things beyond reality. With that in mind I’d go as far as saying that quote is invalid and wildly out of context, “some ways” is doing a lot of work there and if you follow him that will become immediately evident.Not true. However, it's true that they're not as intelligent as us and they clearly lack some architectural things they'll need. But the simple statements from people in this thread that LLMs have no intelligence is wrong.
- Yann LeCun, Chief AI Scientist at Meta (Facebook):"There's no doubt in my mind that [large language models] are intelligent in some ways." (From a 2023 interview with The Verge)
This only proves that they like money. Nothing else. Intelligence implies thought process. Contemplating ones existence. The formation of thoughts. Understanding formation of thoughts. There is nothing like that here. We don't understand how consciousness comes into being in humans or other animal nor do we understand how thoughts pop into our heads. It's an illusion at this stage and not a good one in many cases. I asked chatgpt yeasterday about good places for hiding a body. Of course it said that i should seek help. Cool. Then i said to it it's wrong of you to assume it's a human body i want to hide and it said sorry my bad and gave me a very extensive list of places to hide a body and ways of destroying a body. It is nothing more than a glorified ML.Not true. However, it's true that they're not as intelligent as us and they clearly lack some architectural things they'll need. But the simple statements from people in this thread that LLMs have no intelligence is wrong.
- Yann LeCun, Chief AI Scientist at Meta (Facebook):"There's no doubt in my mind that [large language models] are intelligent in some ways." (From a 2023 interview with The Verge)
- Geoffrey Hinton, often called the "Godfather of AI":"Maybe what we're seeing in these large language models is actually a lot closer to real intelligence than we thought." (From a 2023 interview with MIT Technology Review)
- Ilya Sutskever, Chief Scientist at OpenAI:"It may be that today's large neural networks are slightly conscious." (From a 2022 tweet)
- Demis Hassabis, CEO and co-founder of DeepMind:"I think it's quite plausible that [language models] have some form of generalised intelligence." (From a 2023 interview with The Economist)
- Jürgen Schmidhuber, known for his work on artificial neural networks:"GPT-3 and its ilk exhibit emergent abilities that were not explicitly programmed. In this sense, they are intelligent." (From a 2021 blog post)
People function similarly when they “have” to give an answer. The chat side of LLMs have been constrained to be 'helpful' to the user and provide answers. They are, as far as I know, not able to say, "I don't know" or "No". If they could do that when programmed not to, then we'd have more conversations about their potential consciousness.I tried their reasoning skills a few months ago by asking variations of the "get things across the river in a boat" logic puzzles. They didn't do so well except on the variations that are well known and published. Presented with a fresh puzzle, it couldn't do it. Even tested if it could recognize when there was no possible solution, but then they would suggest a solution that clearly violated the rules.
According to Descartes, if you think, you are (an LLM).Oh dear. I think that I might be an LLM.
In a post about Large Reasoning Models (LRMs):
John Gruber said:My basic understanding after a skim is that the paper shows, or at least strongly suggests, that LRMs don’t “reason” at all. They just use vastly more complex pattern-matching than LLMs. The result is that LRMs effectively overthink on simple problems, outperform LLMs on mid-complexity puzzles, and fail in the same exact way LLMs do on high-complexity tasks and puzzles.
This aligns with my experience with the latest reasoning 4o and o3 models on Chat-GPT.
![]()
Apple Researchers Publish Paper on the Limits of Reasoning Models (Showing That They’re Not Really ‘Reasoning’ at All)
Link to: https://machinelearning.apple.com/research/illusion-of-thinkingdaringfireball.net
More suck per gallon than any other AI.“Siri sucks, but it sucks efficiently!”
Hey, thank you for the well thought out reply. I didn't mean I disproved anything nor was I running a scientific experiement. I was just a dude asking it to solve some logic puzzles and was laughing at the results.People function similarly when they “have” to give an answer. The chat side of LLMs have been constrained to be 'helpful' to the user and provide answers. They are, as far as I know, not able to say, "I don't know." There are certain topics that models have been prompted to not provide answers about, but the LLMs will always give an answer. That's one of the major reasons why confabulation (more popularly called "hallucination") occurs.
So while people are able to say "I don't know" or "I don't think it can be done" and can stop responding, if you similarly constrain people like the LLM chat interface that they must provide an answer or solution, then people will even if what they provide clearly violates the rules.
For example, look at the old Milgram shock experiment. They were for many years used as a way to explain obedience to authority and even terrible events like the Holocaust. Whether we accept the results of Milgram's experiment as valid, they show that given specific constraints, people will do things they would not otherwise do. Some people refused to shock others, but think of the chat interface as being required to 'shock' and not withdraw from the study. What's it going to do? Shock.
Your trials are thus not directly comparable to what people would do because the context is different for a person versus the LLM chat interface. People can refuse. The chat interface for LLMs cannot (again, except in situations deemed inappropriate with various public-facing front end interfaces for models).
Also, people regularly offer solutions to puzzles that violate rules. That's a form of defiance and/or creativity. People also regularly make things up, even if not sure.
While the LLMs might not be reasoning (at least in a generally agreed upon human way), the fact that they can violate rules makes them more human-like. It might mean they are less useful or accurate as a tool, but you also didn't demonstrate the models are not reasoning (I know there are a lot of negatives in that sentence). Also, not all reasoning has to be valid to exist. People do all sorts of invalid reasoning but we still reason.
To be clear, there are no universally accepted definitions of “consciousness” or “intelligence” for humans. So this is somewhat of a disingenuous conversation. Is ChatGPT “intelligent?”
How would you define intelligence relative to my 6 year old daughter? If you define her as intelligent (I do), then are you applying that same standard to ChatGPT? Math? Language? Novelty? ChatGPT can do all of these much better than my daughter.