I have a Macbook Pro M4 MAX that is "maxed" (get it? lol) 128GB RAM and 8TB, fully loaded. I am interested in running some of the advanced LLMs (DeepSeek comes to mind) and I wonder what others have for experience here. I am a "newbie" to running the LLM locally, but I am a regular user of ChatGPT (and I've found that useful).
One thing noted is that, thus far, there doesn't appear to be a way to expand the system with an attached array of GPUs; something the Intel Macs were able to do. I'm going to imagine Apple is well aware of this limitation -- but I'd like to understand more about this, and whether anyone has heard of plans to allow attached GPU arrays to allow and augment better performance.
I think DeepSeek is disrupting enough that it will grant hope that we won't need to spend tons of money on this. And moving forward.
One thing noted is that, thus far, there doesn't appear to be a way to expand the system with an attached array of GPUs; something the Intel Macs were able to do. I'm going to imagine Apple is well aware of this limitation -- but I'd like to understand more about this, and whether anyone has heard of plans to allow attached GPU arrays to allow and augment better performance.
I think DeepSeek is disrupting enough that it will grant hope that we won't need to spend tons of money on this. And moving forward.