Nuff said, are there any MLheads in the forum and potential insights on picking this up instead of a Mac Studio ?
@splifingate mind sharing your workflow ? I have the invoice drafted for two, still on the edge after the reviews outlining weak inference.
As of yet, I don't have a 'workflow': this is just a learning (to understand the mechanisms and paradigms) machine for me.
Initially, I had decided on getting two DGX's (rather than investing in a RTX Pro 6000).
Though I am missing-out on the higher-end networking features (and by-virtue, all the aspects of learning available in this space) by going solo, I have concluded that one is enough for me to learn the majority of what I thought I might need to know.
Depending on your previous experience(s), you may find that the DGX is not gana pull-ahead in inference. Think of it as a whole-package learning machine that excels at more baseline levels. While not particularly exemplary in one (or another) area, it provides a well-rounded, performant environment.
For these (and a few more) I have decided to let my dual-setup reservation lapse.
I can always get another (or more/other), in the future![]()
Thank you for sharing. I feel more at ease letting the reservation lapse when reading the you did the same. Saving up for the M5 studio and getting the DGX next calendar year then.
You may not really need the DGX, but the CUDA integration might tip that scale
Full-Stack seemed important for my introduction into the Arena. I may have more to share after I find the space to work with it more.
I have a few things I want to achieve with VLMS and point clouds, we'll see when I get to it. The to-do list is daunting.
Would love to hear more from your experience.
I'm currently teaching FT, working-through my MAT, and being the GC for my house reno(s) . . . this leaves little time for my inner geek to get out
I will share what I can (when I can).
Considering both right now. Difference in ram vs models able to run is what I am working through now. Hate to say looking at M3 Ultra from China (scary...)Nuff said, are there any MLheads in the forum and potential insights on picking this up instead of a Mac Studio ?
I’ve been doing a lot of that myself recently. I’ve downloaded a lot of models over the months since I got my Studio, and not all models are the same.I am making a bit of a spreadsheet on which (x)billion parameters run on (machine RAM limit). Then going to decide "do I really need to run the 120b vs the 70b?
This is what I am wondering. Only reason to M3 Ultra 512GB would be for deepseek-r1 650b, qwen3:235b, or similar, at this time.Running the larger parameters models is more of a “because you can” rather than “because they’re better”.
Another reason for more RAM is if you want to run more than one AI at the same time.This is what I am wondering. Only reason to M3 Ultra 512GB would be for deepseek-r1 650b, qwen3:235b, or similar, at this time.
One additional thought that may sway you is context length. Many LLMs these days have large context lengths, so can handle larger documents or longer chats without losing track of things. Context length requires RAM.This is what I am wondering. Only reason to M3 Ultra 512GB would be for deepseek-r1 650b, qwen3:235b, or similar, at this time.
Funny, that was my exact thought waking up this morning. I realized that more RAM would get me the headroom to run simultaneous tasks.Another reason for more RAM is if you want to run more than one AI at the same time.
Fortunately, my use-case for the Studio is more hobby-oriented. I love to learn and push tech (remembering back to a ZX81 with a 16K RAM pack that I just *had* to push to its limits!). The “restrictions” of 128GB aren’t really making my workflow “suffer” because there are many alternatives using less RAM.Funny, that was my exact thought waking up this morning. I realized that more RAM would get me the headroom to run simultaneous tasks.
I agree not to overbuy since in 3 years, better will be available. Are you happy with the 128GB? Do you think you will be able to really put it to use for the next 3 years without feeling that somehow your work is really suffering?
![]()
AMD swoops in to help as John Carmack slams Nvidia's $4,000 DGX Spark, says it doesn't hit performance claims, overheats, and maxes out at 100W power draw — developer forums inundated with crashing and shutdown reports
The $4,000 Grace Blackwell dev kit is rated for 240W and 1 PF of sparse FP4 compute, but early users report 100W ceilings and reboot issues under sustained load.www.tomshardware.com
Apparently all is not well in DGX Spark-land... Anyone here with a unit able to corroborate Carmack's claims?