Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.

pethenis

macrumors newbie
Original poster
Mar 28, 2013
7
7
Experts, with drivers for the Mac coming, is the extra money buying a 1080 card wasted, because the 5.1 can't even use the potential of the 1070?
 
Last edited:
1070 is cheaper, uses less power and is almost as fast. Unless you need the extra RAM, the 1070 is probably the better card.
 
I doubt the video card is the issue.... i mean between the slower ram and the now dated (yes still powerful but dated) cpu ... the 1070 would be overkill.
 
It depends on your usage case. Are you going to play games, or use it for professional video work, or something else? For games, you'll probably be CPU limited at anything less than 4K, but if you're going to be using it to run DaVinci Resolve you'll want the biggest GPU you can find.
 
It depends on your usage case. Are you going to play games, or use it for professional video work, or something else? For games, you'll probably be CPU limited at anything less than 4K, but if you're going to be using it to run DaVinci Resolve you'll want the biggest GPU you can find.

Yeah, its all about usage case. Its funny I'll see posts on here on how the cMP can't fully use these GPUs or is bottlenecked cause of other factors. But I render all the time and utilize all 12 cores and full GPU VRAM on my 4K video work. When you work with hi-res footage with multiple effects, color grades, etc., you can use all the power you can get. Even with one effect like NeatVideo, you can maximize all system resources, since it allows you allocate as many cores and GPUs with all VRAM to the effect itself.

But yes, if you aren't doing anything in that realm with your machine, then you have to ask yourself whether an upgrade like this is even needed.
 
Yeah, its all about usage case. Its funny I'll see posts on here on how the cMP can't fully use these GPUs or is bottlenecked cause of other factors. But I render all the time and utilize all 12 cores and full GPU VRAM on my 4K video work. When you work with hi-res footage with multiple effects, color grades, etc., you can use all the power you can get. Even with one effect like NeatVideo, you can maximize all system resources, since it allows you allocate as many cores and GPUs with all VRAM to the effect itself.

But yes, if you aren't doing anything in that realm with your machine, then you have to ask yourself whether an upgrade like this is even needed.

I edit video (Some 4K) in Premiere, and colorgrading in Davinci Resolve on a 4.1 -> 5.1 12 core 3.46 with 48 Gig ram and SSD's.
 
I edit video (Some 4K) in Premiere, and colorgrading in Davinci Resolve on a 4.1 -> 5.1 12 core 3.46 with 48 Gig ram and SSD's.

We have very similar systems. I currently use the Titan X (Maxwell) to great effect. A 1080 should serve you well, assuming the drivers are fine.
 
Is this official? Could i pack me a 1080 and run? Or does this require flashing stuff and other modifications?
 
It depends on your usage case. ...

Yeah, its all about usage case. ...

We have very similar systems. I currently use the Titan X (Maxwell) to great effect. A 1080 should serve you well, assuming the drivers are fine.
Thanks for helping to combat the simple FUD that a cMP can't fully utilize a high end GPU.

While there certainly are applications where the CPU is a bottleneck, and perhaps many games are in that category, the softwares using CUDA GPGPU acceleration often are able to saturate the GPUs without putting great load on the CPUs.

Even PCIe bus issues with the older systems can be unimportant for GPGPU apps that spend most of their time crunching on stuff in VRAM, and not so much time moving data between system RAM and VRAM. (And Pascal's unified memory helps with that issue.)
 
Last edited:
Is this official? Could i pack me a 1080 and run? Or does this require flashing stuff and other modifications?

Answered your question in another thread where you asked the same question:

To fully answer your question. As soon as the proper Web Drivers are available you will be able to use a GTX 1080 just like any other Nvidia card with the current web drivers. As far as a boot screen, an EFI flash will be available from MacVidCards. You will be able to buy the card directly from him or buy you own card forward it to him and have him flash it for you.

Lou
 
  • Like
Reactions: pat500000
GTX 1080 will be Bottlenecked by your CPUs, regardless of workload you will post on it. Go for GTX 1070. Its fast enough, uses less power, and is cheaper.
 
GTX 1080 will be Bottlenecked by your CPUs, regardless of workload you will post on it. Go for GTX 1070. Its fast enough, uses less power, and is cheaper.

This is simply not true. There are plenty of "pro" workloads like DaVinci Resolve where the GPU will be crunching away on the data with no interaction with the CPU (or PCIe bus). For cases like that, you want the biggest GPU you can get your hands on.
 
GTX 1080 will be Bottlenecked by your CPUs, regardless of workload you will post on it. Go for GTX 1070. Its fast enough, uses less power, and is cheaper.
This is simply not true. There are plenty of "pro" workloads like ML training where the GPU will be crunching away on the data with no interaction with the CPU (or PCIe bus). For cases like that, you want the biggest GPU with the most VRAM that you can get your hands on.

ML training workstation:
4xTitans.jpg

24 CPU cores, 512 GiB CPU RAM, 14,336 CUDA cores, 48 GiB VRAM.

GPUs pegged, CPUs mostly idle.
 
Last edited:
  • Like
Reactions: zoltm and h9826790
This is simply not true. There are plenty of "pro" workloads like DaVinci Resolve where the GPU will be crunching away on the data with no interaction with the CPU (or PCIe bus). For cases like that, you want the biggest GPU you can get your hands on.
This is simply not true. There are plenty of "pro" workloads like ML training where the GPU will be crunching away on the data with no interaction with the CPU (or PCIe bus). For cases like that, you want the biggest GPU with the most VRAM that you can get your hands on.
The OP posted his workload. Will he not be CPU bottlenecked?

He will be. Simple as it can be.

Why would I talk about a situation, that is out of context of the thread? In his context HE WILL BE BOTTLENECKED REGARDLESS OF WORKLOAD HE WILL PUT ON THE GPU.

If the OP wants, and can afford spending 150$ more, for 10% more performance vs GTX 1070, its up to him to decide. If it would be my money, I would go for either GTX 980 Ti, or GTX 1070. More likely GTX 1070, because it will be much easier to set up in MP 5.1, because there are models that have only 6 pin connector(I think Founders Edition has 6 pin connector, but I can be wrong on this).
 
Why would I talk about a situation, that is out of context of the thread?
On the interwebs, however, few people take the time to read the entire thread so that a simple declarative statement is often taken at face value, rather than being understood in the context of all that went before. It is useful for everyone to repeat a bit of the context before a simple, abosolute statement.

And note that in post #7 the OP says that he uses DaVinci Resolve, and in post #31 it was stated that DaVinci Resolve is GPU bound.

So, before accusing others of not following the context, you should pay attention to the context.
 
  • Like
Reactions: h9826790
On the interwebs, however, few people take the time to read the entire thread so that a simple declarative statement is often taken at face value, rather than being understood in the context of all that went before. It is useful for everyone to repeat a bit of the context before a simple, abosolute statement.

And note that in post #7 the OP says that he uses DaVinci Resolve, and in post #31 it was stated that DaVinci Resolve is GPU bound.

So, before accusing others of not following the context, you should pay attention to the context.
OP said about specific Workload in DaVinci resolve. Will he be in that workload GPU bound 100%?

No. So in his case, he will sometimes rely on CPU performance, and therefore - bottlenecked by it.
 
OP said about specific Workload in DaVinci resolve. Will he be in that workload GPU bound 100%?

No. So in his case, he will sometimes rely on CPU performance, and therefore - bottlenecked by it.
But if sometimes she is not bound by CPU, then the 1080 could be a better choice.

So, the answer depends on what "sometimes" means. If she's constrained by CPU 90% of the time, then a cheaper GPU might be more cost effective. If 90% of the time the GPU is her bottlenect, then she should get the fastest GPU that she can find.

Your Nvidia hate blinds you to looking at the true situation and the need to balance the pros and cons.
 
  • Like
Reactions: h9826790
But if sometimes she is not bound by CPU, then the 1080 could be a better choice.

So, the answer depends on what "sometimes" means. If she's constrained by CPU 90% of the time, then a cheaper GPU might be more cost effective. If 90% of the time the GPU is her bottlenect, then she should get the fastest GPU that she can find.

Your Nvidia hate blinds you to looking at the true situation and the need to balance the pros and cons.
Right, because I am saying that in his/her particular scenario GTX 1080 will be bottlenecked It shows my hate for Nvidia. Aiden. I guess it depends on what YOU want to see in my posts.

I know all of what you are writing. But think about a little about the particular scenario. Will it be bottlenecked? Yes. 90% of the time. Is it worth to pay 150$ more for those situations? No. End of the story. I do not know, how the f*** you can come up with idea that I am hating Nvidia this way. Maybe its not me who is blind?
 
4K video editing and DaVinci Resolve are definitely GPU-limited tasks, even on a cMP. If he was planning on just playing games, I'd definitely agree with you that he would be CPU limited for the majority of cases. I'm not sure why you think 4K video editing with Adobe products that use CUDA would be CPU limited?
 
Maybe you are right guys, that in this scenario he will not be bottlenecked. I guess, we would have to experiment with it, to know what is happening, exactly.
 
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.