Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.

VitoBotta

macrumors 65816
Original poster
I used to play around with LLMs on my M4 Pro mini last year, but I kind of stopped because I only have 64 GB of RAM. That meant I could only run small models, and those just aren't very useful compared to the hosted ones. What about you? Are you actually using LLMs on your Mac for real work? If yes, which models are you using?
 
I've tried a couple of times on my M1 Pro 16GB RAM but got the same results, the best LLM it can run is minimax-2.5 with a "not so low" response time on Ollama, bigger ones, like kimi-k2.5 take forever to answer.
 
Not intentionally, but often I ask Siri something and she goes "That's too hard, kin I ask ChatGPT?"
 
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.