I gotta hand it to Dave Lee
He's made a career on YouTube from playing with gadgets while usually looking like like he just rolled out of bed
😂
Would it be safe to say that in 5-10 years a smartphone will be able to run a model like this internally and without the internet?
I highly doubt that the typical memory size in a smartphone is going to jump from 8-12gb to 512gb-1tb in the next 5 years, and the only way I even see it in 10 is if we see a move to some kind of optane-like unification of memory and storage, which is unlikelyWould it be safe to say that in 5-10 years a smartphone will be able to run a model like this internally and without the internet?
We are entering a new era of computing where VRAM is becoming increasingly important. Apple's past slow trends in RAM upgrades are not a reliable guide for the future.While early years saw some doublings of the RAM (iPhone 3GS, iPhone 4, iPhone 5, iPhone 6S), those have become rarer, and RAM increases have also become smaller (e.g., from 6 GiB to 8, rather than to 12).
By 2035, maybe we'll have an iPhone with 64 GiB RAM. Almost certainly not 512.
Storage in the Apple Silicon era is getting faster almost annually. We may soon see a unified memory / storage architecture that treats RAM essentially as cache while leaving data in place on the SSD and treating it as it were in RAM.No smartphone will have 512GB of RAM in 10 years. At least not from Apple, if we can go by history. Unless 32GB of iPhone RAM is analogous to 512GB of Mac Studio RAM
Can somebody explain the actual use of an LLM that cannot search for up-to-date info online?
I'm not being sarcastic, I actually want to know.
Wouldn't the knowledge pool of a local solution be limited by its knowledge at the time of release?
Thanks; I'm really not an expert in this field.
Not iPhones as they’ll ship with 8GB memory.Would it be safe to say that in 5-10 years a smartphone will be able to run a model like this internally and without the internet?
I do not know how Llama but Deep seek main core is just 32B and rest are other specialized agents loaded into memory as well to be available instantly. That is why 70B model runs slower then this.Probably sooner if you mean equivalent performance, not model size. You can already run Google Gemma 3 27B or Qwen QwQ 34B on a spect out MBP, and they are close to Deepseek in performance.
Just last year Meta’s Llama 3.3 70B matched the performance of Llama 3.1 405B released 6 months prior. X6 efficiency improvement! Llama 3.1 405 was itself close to original GPT4 which had many trillions of parameters. From needing 8 Nvidia h100 to one MBP in two years.
And, I think part of the delay of bringing the personal features to the market is due to the changing political climate where certain countries don’t want some personal data (that’s required for things like this to work) to be automatically available to the user. Just saw a law passed in Germany where Google’s Calendar app can’t surface birthdays in contacts. If a user wants to have birthdays, they need to enter them manually. This can add another layer of complexity that wasn’t there before and would need a rethink before it could be offered.With all the crying going on I think Apple is doing this exactly right.
I wonder how good this thing will do Video Ai stuff. Like Topaz new Starlight or Video Ai work.
More importantly, can it run Doom?Cool - but how fast can it load MacRumors.com in Safari??? I want real-world cases for myself before I plunk down $15K.
Unclear. In terms of technology the biggest problem is probably the DRAM.Would it be safe to say that in 5-10 years a smartphone will be able to run a model like this internally and without the internet?
We are entering a new era of computing where VRAM is becoming increasingly important. Apple's past slow trends in RAM upgrades are not a reliable guide for the future.
But Apple Silicon 2 will mean 64 GB of RAM will be like 512 GB of Apple Silicon 1 RAMThey can run small models today but it doesn’t seem likely that they’d scale to 512GB of RAM in 10 years.
So 5-10 years from now they will be able to run larger models than the small LLMs we can run today, but no not this sized model.
We don't quantify it, it just needs to feel snappier!Cool - but how fast can it load MacRumors.com in Safari??? I want real-world cases for myself before I plunk down $15K.