Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.
From IBM's Power4 Website -- 32/64 bit compatibility

Here is what IBM says about the processer:

Maintain binary compatibility for both 32-bit and 64-bit applications with prior PowerPC and PowerPCAS systems: Several internal IBM task forces in the first half of the 1990s had concluded that the PowerPC architecture did not have any technical impediments to allow it to scale up to significantly higher frequencies with excellent performance. With no technical reason to change, in order to keep our customers software investment in tact, we accepted the absolute requirement of maintaining binary compatibility for both 32-bit and 64-bit applications, from a hardware perspective.

See for yourself:
http://www-1.ibm.com/servers/eserver/pseries/hardware/whitepapers/power4.html
 
Originally posted by barkmonster

At the time G3s were at 266Mhz and the Pentium II and K6 cpus were at 600Mhz.

There wasn't ever a time when 266MHz G3s competed against Pentium IIs and K6s at 600MHz. The only desktop that the 266MHz G3 shipped in was the Beige tower and the iMac (iMac may be the only exception to my argument here). The Beige G3 competed against the PII 233-333MHz machines. Apple moved to the 100MHz bus right after Intel and the 300+ MHz Blue and White G3s competed with the 400+ MHz PIIs (though slower PIIs, LX based machines, were still being sold for a while).
Aside from this, the PII NEVER ran faster than 450MHz anyway. The PIII came to market at 450MHz, replacing the PII.
The K6/2 ran at similar speeds during the processor race, though sometimes clocking slightly faster. It was WAY easier to overclock since Intel started locking the multipliers on the Pentium cartridges. The K6/2 stayed in production far longer than the PII since AMD didn't have a moble version of the Athlon for a LONG time. This is why K6/2s reached higher clock speeds... up to 550MHz, though they didn't get that fast until after the K7 was available.

Saying that and being WAY too optimistic, when a next generation powermac comes out and intel keep using design compromises to push their clockspeeds higher we could be in a situation of apple running those snail ads again by this time next year.

Imagine it :

5Ghz Pentium 4, 1 instruction per clockcycle, huge power requirements and elaborate cooling methods.

2Ghz PowerPC, 8 instructions per clockcycle, low power and turbine cooled. [/B]

I'd say that IS way too optimistic. The P4 can execute 3 operations per clock cycle. It may average slightly over 1 operation/clock but it can do more.

The PowerPC64 platform is supposed to be 8-way superscaler though no one seems clear on what IBM means by that (I'm not a CPU designer). The definition of superscaler that I've seen simply say that superscaler processors are designed to run multiple instruction at once. I've not seen any wording indicating that x-way superscaler means it can run x operations per clock cycle. I've also not seen any wording that says it can run a sub-set of x possible operations per cycle.

Look at the current G4 for example. It is superscaler and it can issue 3 or 4 operations max per clock cycle. It can issue one operation in any one of 8 area's fo the CPU [load/fetch, integer, fp, fp, altivec, altivec, altivec, altivec]. Does this mean the a G4 is 3 way superscaler or is it 8 way superscaler? I've heard both arguments and I honestly don't know which is true.

If you consider both definitions you'll realize that definition one, the 8 instructions/clock, would require a MASSIVE processor to pull off. There would be multiple load/fetch units, multiple integer, multiple fp, and altivec on the chip. It'd be a monster. This would mean that it would be very big, very hot, and very expensive. This isn't the ideal chip for Apple... It'd never go in a laptop.

If you consider definition two, the 8 instruction units with a subset of instructions/clock ala the current G4, that would be much easer to pull off. The chip size would be manageable. It could be made to run more efficiently than a G4, higher average # of instructions/cycle. It would be more powerful as a result. It won't run 8X as many instructions/cycle as a P4 though.

I think that IBM's next chip will be able to issue a subset of 8 possible instructions per clock cycle. We won't know for sure until next week though. We also won't know what the average number of instructions/cycle actually is for quite a while.

ffakr.
 
Re: Re: Re: This just adss fuel to the fire

Originally posted by MacCoaster

Uh. The entire M68k family is 32bit, even the original Motorola 68000.

Right, but the early ones only had 24-bit addressing, and some programs used the upper eight bits of the address word for things that would be very bad if the entire 32-bit word was being interpreted as an address.

As to the PowerPC, there should be no need to change the OS at all. The PowerPC was always designed to be implementable as either a 32- or 64-bit chip, with 64-bit implementations being backwards compatible with 32-bit code.
 
Even if Apple has to sacrifice chip design to get ultra high megahertz they need to do it. If you want to sell computers you have to have the most megahertz. Alti-vec...and all this other stuff Apple has been feeding us will not improve speed only more Megahertz actually makes a faster computer. I love my Macs but am thinking PC more and more because they are SO much more powerful right now. It costs thousands more for a 1.25 dual then what you can get a cheap PC for that is going to run a lot faster. Heck I even saw a $399 PC with a 1.7 GHZ processor in it...that has more speed and will kick the **** out of a PowerMac.

I love my Apples, but am afraid that they are not as good of values as they used to be. And also am afraid that this new IBM chip will actually be slower then the current ones, since according to many on these boards, Apples have actually been becoming slower and slower over each revision. I have thought about replacing my iMac DV 400 with a new eMac 800, but from what people are saying it does not sound like it will be any faster.

I have never used a PC and have always enjoyed Macs, but acording to even most Mac people PC's and XP is better now. Too bad, I like supporting the little company.

When I see a $399 PC with 1.7 GHz and a $3,500+ Mac with only 1.25 GHz, then I can see these peoples points.

I guess I am just getting down on Apple, because other people are.
 
Originally posted by Abercrombieboy
Even if Apple has to sacrifice chip design to get ultra high megahertz they need to do it. If you want to sell computers you have to have the most megahertz. Alti-vec...and all this other stuff Apple has been feeding us will not improve speed only more Megahertz actually makes a faster computer. I love my Macs but am thinking PC more and more because they are SO much more powerful right now. It costs thousands more for a 1.25 dual then what you can get a cheap PC for that is going to run a lot faster. Heck I even saw a $399 PC with a 1.7 GHZ processor in it...that has more speed and will kick the **** out of a PowerMac.

I love my Apples, but am afraid that they are not as good of values as they used to be. And also am afraid that this new IBM chip will actually be slower then the current ones, since according to many on these boards, Apples have actually been becoming slower and slower over each revision. I have thought about replacing my iMac DV 400 with a new eMac 800, but from what people are saying it does not sound like it will be any faster.

I have never used a PC and have always enjoyed Macs, but acording to even most Mac people PC's and XP is better now. Too bad, I like supporting the little company.

When I see a $399 PC with 1.7 GHz and a $3,500+ Mac with only 1.25 GHz, then I can see these peoples points.

I guess I am just getting down on Apple, because other people are.


Whoaaaaaaaaaaaaa! Slow down there buddy! :D

1) $399 PC, I won't touch one with a stick, unless I am building it and I got some serious deals on the parts, and I am willing to make major sacrifices.
2) $399 PC will not be faster than your powermac, it will be cheaper, and will have smaller HD, slower components and everything built into the mother board.
3) If you have never used a PC, I don't think you should do a reverse-switch. Most people I know who did that miss their old days when they were Mac users and are too much invested in their current envirornment to switch back.
4) The eMac will smoke your iMac DV 400. If you go by megahertz as you like to believe the speed of a system is measured, then it will be twice as fast as your iMac and would probably cost you less than you paid for your iMac.

Don't believe everything you read, and you DO get what you pay for.

my 2.5 cents.
:D

P.S. You will pay way more than $399 to get it to pefform the way you want it to, and by that time your computing experience would go down the dell ... errr ... toilet. :p
 
Originally posted by Abercrombieboy
Even if Apple has to sacrifice chip design to get ultra high megahertz they need to do it. If you want to sell computers you have to have the most megahertz. Alti-vec...and all this other stuff Apple has been feeding us will not improve speed only more Megahertz actually makes a faster computer.

The product of Intel marketing? :eek: You decide. ;)
 
Originally posted by Abercrombieboy
I love my Macs but am thinking PC more and more because they are SO much more powerful right now. It costs thousands more for a 1.25 dual then what you can get a cheap PC for that is going to run a lot faster. Heck I even saw a $399 PC with a 1.7 GHZ processor in it...that has more speed and will kick the **** out of a PowerMac.
I'm the first to admit that a higher clocked processor is probably going to be faster in most common tasks... a 2GHz P4 will run a bubble sort (or even Netscape) faster than a 1GHz G4.
You are totally on a different page though. You are comparing a 1.7GHz celeron with almost no L2 cache to a a Pro line Mac tower. I guarantee your $399 tower is going to have crap video, crap hard drives, crap sound, and if it even has a NIC.. that is going to be crap too. You're comparing an Escort to a Mercedes. Apple computers are made from very high quality components. There is a reason why I see 8 year old macs still running every day on my campus.
I love my Apples, but am afraid that they are not as good of values as they used to be. And also am afraid that this new IBM chip will actually be slower then the current ones, since according to many on these boards, Apples have actually been becoming slower and slower over each revision. I have thought about replacing my iMac DV 400 with a new eMac 800, but from what people are saying it does not sound like it will be any faster.
I'm not sure where you get these notions from. Apple were NEVER good values. You pay for the user experience, and you pay for machines that are built better. You do pay though, and you always have.

The new IBM chips will NOT be slower... I'll guarantee that. They are built using the Power4 processor as their design reference. The Power4 (though ONLY clocked at slightly over 1GHz right now) is the fastest computer chip in production today. It toasts G4s, P4s, Athlons, Itaniums, HP/RISC (though not by much any more)... Strange how a chip that only runs at 1GHz could be so fast isn't it?? Guess there is more to processor design that speed at any cost.

Also, why would you think that an eMac 800 won't be any faster than an iMac DV 400? The processor is twice as fast (clocked), it has altivec support, it has faster drives, the eMacs video is at least two generations newer [faster] than the iMacDV's video. The eMac isn't going to impress someone who uses a 2.4 GHz P4, but it is a nice machine (I've used them quite a bit), and it is marketed at the low end educational buyer and more recently the low end consumer.
I have never used a PC and have always enjoyed Macs, but acording to even most Mac people PC's and XP is better now. Too bad, I like supporting the little company.
Most people say that XP is much better than other versions of Windows, but most reviews of XP vs. OS X (even by PC magazines) give the edge to OS X. I have seen a LOT of reviews that basically say that XPs user experience is a poor copy of OS X's. I'd, for the most part, agree with that.. though there are things about Windows that I prefer.
As for supporting the company, that is admirable... especially since Apple has always driven innovation in the computer industry. MS would be selling OSes that were still DOS shells if other people didn't push them to make better products. I'm a realist though. I always tell people to get the best product for their needs. If you can't afford a Mac, get a PC. If you have used PCs for years with out major issues, stick with what you know. This isn't to say that I don't think that the Mac is a better tool overall.
When I see a $399 PC with 1.7 GHz and a $3,500+ Mac with only 1.25 GHz, then I can see these peoples points.
Again, you are comparing a single processor Celeron to a dual processor 1.25 GHz G4 tower. You are comparing a bargan bones PC made with the cheepest components to a Pro Mac. I'd wager that if you purchased both and used them equally, only one computer will be useful in 2 years... only one computer will still be running in 3 years. Guess which one?

...just my 2cents.
I am, after all, just a stupid ffakr.
 
Re: Re: Re: This just adss fuel to the fire

Originally posted by gropo
Yes, THANK YOU... Finally someone who knows what's up! I read over the dev docs @ the ADC regarding the G4 fpu addressing protocol - 32-bit fpu calls are referred to as "single" and 64-bit calls are referred to as "double," indicating that the groundwork has already been laid for the advent of 64-bit integer register space.

i think this might be slightly confused. 32 bit floating point values are traditionally called single precision floats (or "single" or even just "float" in C) whereas a 64 bit floating point value is a double precision floating point value (or "double" in C). Support for both single and double precision FP has been common for ages. It is independent of how many bits are present in integer/address registers.

That said, Apple appears to have been very careful to ensure "type opacity" for pointer values, so they can easily change from being a 32 bit quantity to a 64 bit quantity with not much more than a recompile operation.
 
People keep saying the POWER4 can manage 8 instructions/cycle, it can't. It can fetch 8 and dispatch up to 5 (1 group).
 
Re: Re: Re: Re: Re: Re: This just adss fuel to the fire

Originally posted by pianojoe

If a program writes out a memory location to a file, the 32bit version will write out 4 bytes, and the 64bit version will write out 8 bytes. The data files will not be compatible without extra effort.

What you say is accurate, but misses the point. Each program should write 32 bits. It's NOT a problem to define data quanties as being 8, 16, 32 or 64 bit independent of the native integer/pointer register size, provided you've not been plain lazy/dumb at the start.

Or put another way, a 64 bit processor can still read/write images (ETC) that a 32 bit processor can read/write and vice versa (etc etc). Otherwise it would really be an incompetent mess!

If there are any issues encountered with specific programs here, the original authors didn't deserve a salary.
 
Ffkar,

When they say that this new processor chip will be a 8 way supercalar it means that it can perform 8 indepent instructions per clock cycle. To do this, the processor must have an instruction fetching unit that can fetch more than one instruction at a time, built-in logic to determine if instructions are independent, and multiple execution units to execute all the independent instructions. Of course, that instrucions that depend on the result of the previous instruction can't be processed at the same time. You cant have the chicken and the egg at the same time. First the chicken, then the egg.
The current G4 is a 4 way Superscalar chip.
Imagine this: You have an instruction (a instruction that takes advantage of the superscalar technology) that you would process on both Chips (a 4 way and a 8 way superscalar). The 4 way superscalar was a 1Ghz and the 8 was a 500Mhz... they would both dispatch the instruction at the same time! That's the benefit of the Superscalar technology.
 
Originally posted by ffakr

The new IBM chips will NOT be slower... I'll guarantee that. They are built using the Power4 processor as their design reference. The Power4 (though ONLY clocked at slightly over 1GHz right now) is the fastest computer chip in production today. It toasts G4s, P4s, Athlons, Itaniums, HP/RISC (though not by much any more)...

Not to be argumentative, but this simply isn't true. I totally dig the concept of Macs blowing the doors off Windows machines, but it's simply not going to happen. There's too much inertia on the side of Intel/AMD right now.

I think it's more reasonable to assume these chips (meaning, the IBM GPUL) will benchmark close to current Intels. Look at these SPECfp2000 numbers.

The Itanium is beating the Power4, though not by tons.

And the important thing to remember is that the new IBM chip IS NOT A POWER4. It is a new chip which takes design cues from the Power4. It is being referred to as GPUL (GigaProcessor UltraLite), which is a reference to it being the "light" version of the Power4.

Basically, that means they scaled it back from the heat-producing electrical black hole it is now into a more reasonable chip that doesn't require you to also build your own nuclear reactor to run it. Then they added on an AltiVec-alike, and appear to be extending the pipeline in order to boost Mhz performance, probably to make it easier to market to the average consumer.

(I don't understand why people on this list constantly use the current Power4 numbers...1-1.3Ghz, when IBM has, itself, said that they expect the chip to DEBUT at 2 Ghz. To me, this means that this processor has a longer pipeline, and will not match the Power4 in terms of operating efficiency.)

Undoubtedly, this chip will be fast, but not as fast as the Power4.

Yes, the Itanium stumbled out of the gate. Yes, moving to the Itanium is going to be a bitch for the PC world, because of addressing incompatibilities. Yes, the PowerPC move to 64 bit should be easy.

But no, the GPUL is not going to "toast" the Itanium in raw performance...the Itanium is pretty bad-assed.

(As a side note, the Itanium ALSO runs at slower clock speeds...1-1.5 Ghz. Intel's going to have to completely shift its marketing strategy if they plan to mainstream Itaniums after they milk the P4.)
 
Originally posted by suzerain
And the important thing to remember is that the new IBM chip IS NOT A POWER4. It is a new chip which takes design cues from the Power4. It is being referred to as GPUL (GigaProcessor UltraLite), which is a reference to it being the "light" version of the Power4.

(I don't understand why people on this list constantly use the current Power4 numbers...1-1.3Ghz, when IBM has, itself, said that they expect the chip to DEBUT at 2 Ghz. To me, this means that this processor has a longer pipeline, and will not match the Power4 in terms of operating efficiency.)

You forget, the GPUL will be
1) smaller (no second core, ?less L2 cache ondie?)
2) made on a finer (.13 or .09) process, the current POWER4's are .18.. and frickin' HUGE as a result.. something in the region of 400mm^2

a much smaller die due to not having a second core, perhaps a bit less L2 cache, and a smaller process.. and it could well be the same pipeline length as the POWER4, less signal propogation issues could alone add a fair bit of headroom.

infact, I'd put money (but not much) on it being a POWER4 design with the second core chopped and altivec/VMX execution units added.

and a lone POWER4 core @ 2Ghz is (theoretically) a very fearsome piece of silicon indeed.

Now, the memory bandwidth will be scaled back from the POWER4, and I'd imagine L3 cache for it in Macs won't top 4MB or so, so it'll perform "differently" to a real POWER4, slower on _large_ bandwidth bound workloads, but possibly faster on stuff that'll fit in the L1/L2/L3? caches. (see 500Mhz Celeron/Mendicino verses 400Mhz Pentium 2/Deutches as an x86 example of that behaviour.)
 
Originally posted by suzerain
Not to be argumentative, but this simply isn't true. I totally dig the concept of Macs blowing the doors off Windows machines, but it's simply not going to happen. There's too much inertia on the side of Intel/AMD right now.

I think it's more reasonable to assume these chips (meaning, the IBM GPUL) will benchmark close to current Intels. Look at these SPECfp2000 numbers.
I can't get to that link for some reason. It's OK to be argumentative too... it's a rumor discussion list :)
And... if you look back at my original statement (which you quoted), I was refering to the contention of the previous poster.. that the PowerPC64 will be slower than the G4. I said that I guaranteed that whatever IBM has up it's sleeve won't be slower than the G4. You twisted what I said. :)

The Itanium is beating the Power4, though not by tons.
It wasn't the last time I looked, but I haven't checked out the SPEC on Itanium2 yet.

And the important thing to remember is that the new IBM chip IS NOT A POWER4. It is a new chip which takes design cues from the Power4. It is being referred to as GPUL (GigaProcessor UltraLite), which is a reference to it being the "light" version of the Power4.

Basically, that means they scaled it back from the heat-producing electrical black hole it is now into a more reasonable chip that doesn't require you to also build your own nuclear reactor to run it. Then they added on an AltiVec-alike, and appear to be extending the pipeline in order to boost Mhz performance, probably to make it easier to market to the average consumer.

(I don't understand why people on this list constantly use the current Power4 numbers...1-1.3Ghz, when IBM has, itself, said that they expect the chip to DEBUT at 2 Ghz. To me, this means that this processor has a longer pipeline, and will not match the Power4 in terms of operating efficiency.)
I don't know either. I didn't use 1 GHz in reference to the upcomming PowerPC64, I only used that number to point out that a processor (the Power4 in my example) can be very very powerful with out being clocked very very fast.

It may have a longer pipeline, though IBM was never a big fan of long pipelines in the past. evne the latest PPC750s don't have very long pipes... I think they are like 7 stages in the 750fx but don't quote me on that. As far as I know, pipeline is not the only element affecting clock speed. It's a way to boost it, but the overall design is going to affect how fast the chip clocks too.
Undoubtedly, this chip will be fast, but not as fast as the Power4.
I don't think it will be as fast as the Power4 either. I don't think I ever said that. It won't be competing against the Power4 though, just like it won't be competing against the Itanium or the PA-RISC (which I call the wrong name earlier :-(
This is a small server/workstation processor. Merced (and the Power4) are still being marketed as big iron, or at least mini-big iron.. if that makes sense.
I do, however, think it will be fast. Really fast. I think it will be a smoking processor because:
1) It takes design cues from the Power4, but it will build on the Power4 design. It isn't unreasonable to expect that next gen processors benefit from lessons learned on the last project. Sure it will be a 'lite' processor, but it will benefit from all that IBM has learned from the Power4 (and the power3, and the 750s, and the 604... and so on).
2) If it can really do multiple fetch/stores and issue as many ops per cycle as people say, it will be much more efficient than the x86 processors it will compete with. I recall seeing that the P4 really doesn't average much more than one operation per cycle in the real world. If you put a PowerPC that could _potentially_ execute 3 or 4 times as many operations per cycle as a P4 that was clocked twice as fast as the PowerPC... guess which processor will perform better?
3. it will retain Altivec which has always seemed to perform better than SSE2. Many of the SSE2 instructions take MANY more cycles to complete than the same or similar operation in Altivec.
Yes, the Itanium stumbled out of the gate. Yes, moving to the Itanium is going to be a bitch for the PC world, because of addressing incompatibilities. Yes, the PowerPC move to 64 bit should be easy.
It's not addressing incompatabilities that will hold up acceptance of EPIC. EPIC is totally different from CISC (x86 in this case). Itaniums are VLIW processors. They are HUGE, hot, and they require exquisitely complex compilers to write fast code. Itanium is probably the most complex processor ever created and many people think that the compiler writers will need to work harder than the chip engineers to make Itanium work well.
But no, the GPUL is not going to "toast" the Itanium in raw performance...the Itanium is pretty bad-assed.

(As a side note, the Itanium ALSO runs at slower clock speeds...1-1.5 Ghz. Intel's going to have to completely shift its marketing strategy if they plan to mainstream Itaniums after they milk the P4.)
Itanium doesn't run at 1.5 GHz. I'm pretty sure Itanium 2 is only available at ~900 and 1000MHz so far.

But who cares if it toasts the Itanium. I didn't say GPUL will toast itanium.. I have no Idea about that. I said that the power4 toasts Itanium **which as you tried to point out, may not be true anymore**
If GPUL ends up in macs next year, it won't compete with itanium. I don't expect to see Itanium in the average computer until well after Deerfield arrives. I think the software will hold up acceptance of Itanium at least as much as the hardware will. Intell will push x86 at consumers for AT LEAST two more years, probably longer.
PowerPC64 will go up against P4 and Athlon. It will likely also go head to head with clawhammer. This is where the comparisons need to be made. From what we think we all know, I think it will take them all. clawhammer is the real unknown. It has only sampled at 800MHz so far so who knows how it will perform when it comes out of the gate (about mid way through 2003).
If the PowerPC can do as much work as people claim it can (per cycle), I think it is going to put a lot of PCs to shame.

This is all, of course, only my opinion. Afterall, what do I know, I'm just an average ffakr with the same sketchy info and innuendo that everyone else has. :)
 
Originally posted by Chryx


You forget, the GPUL will be
1) smaller (no second core, ?less L2 cache ondie?)
2) made on a finer (.13 or .09) process, the current POWER4's are .18.. and frickin' HUGE as a result.. something in the region of 400mm^2

a much smaller die due to not having a second core, perhaps a bit less L2 cache, and a smaller process.. and it could well be the same pipeline length as the POWER4, less signal propogation issues could alone add a fair bit of headroom.

So, what sort of relationship is there between the manufacturing process and speed? Does it have a greater impact on speed of the chip, or power consumption? Or is it one of those situations where the designers have to choose where to take their advantage (i.e., going to a smaller manufacturing process allows you to choose either lower power consumption or added speed or some combination thereof)?

I ask, because is it reasonable to assume that going from a .18 micron process to .13 (or .09) will automatically allow them to boost speeds from 1.3Ghz to 2.0 Ghz without extending the pipeline?

I'm just wondering if you could illuminate on that a bit...on how much speed such a change in manufacturing would allow you to gain (all theoretically, of course, since this whole discussion is basically about vaporware, until we see IBM talk about it).
 
Originally posted by ffakr

This is all, of course, only my opinion. Afterall, what do I know, I'm just an average ffakr with the same sketchy info and innuendo that everyone else has. :)

Yeah, well, this is all our opinions, since we have no real data, save for cryptic messages about a product announcement.

However, I didn't "twist" anything you said; I just quoted it. :p

My real point could have been summed up this way:

IBM has said [paraphrasing] that they are going to discuss a new chip which is based on the Power4, but it is destined for desktop workstations rather than servers.

To me, "based on" means, pretty specifically, that it is NOT a Power 4. Otherwise they would have just said, "We're announcing a new version of the Power4 for workstations."

It's obviously different enough that they bothered to make that distinction.

Therefore, accepting all of that logic, it doesn't seem to me that it's particularly helpful to look at Power4 benchmarks AT ALL, since the Power4 is a big fat power-wasting behemoth of a chip that is used in $15,000 servers.

Obviously, the GPUl will be a very different beast indeed. And, as the other poster who responded to my post said, it may well be true that the chip outperforms the Power4 for smaller operations, but would probably be toasted by it on applications where it has to process gigantic amounts of data.

Then again, maybe not. But how would IBM continue to sell servers based on the Power4 if they had a cheaper chip that outperformed it?

So, since the Itanium 2 is edging the Power4 on benchmarks, I think it's also doubtful that the GPUL will outbench that chip, either.

Instead, I'd guess that we'll see performance similar to a Pentium4, but with better FPU perfformance.

And, IMO it's also unlikely that it'll blow the doors off Intel...notice how the whole industry is considering a 64 bit leap at the same time? I think these developments all kinda happen together...

It would seem reasonable to assume, though, that a 2 Ghz GPUL would be able to be considered an equal to a 3-4 Ghz Pentium4, which is likely what'll be on the other side of the fence at the time.

Of course, I'd like it to be...oh I dunno...1,000 TIMES faster than the then-fastest P4...
 
Originally posted by suzerain


So, what sort of relationship is there between the manufacturing process and speed? Does it have a greater impact on speed of the chip, or power consumption? Or is it one of those situations where the designers have to choose where to take their advantage (i.e., going to a smaller manufacturing process allows you to choose either lower power consumption or added speed or some combination thereof)?

I ask, because is it reasonable to assume that going from a .18 micron process to .13 (or .09) will automatically allow them to boost speeds from 1.3Ghz to 2.0 Ghz without extending the pipeline?

Well, I'll give you an x86 example (since their processors are the ones I'm most familiar with, my knowledge of which PPC chips were which process is kinda flaky)

the "Coppermine" Pentium 3, on .18, topped out at ~1.05ghz (that's as high as overclockers could push them in a stable manner with anything other than EXTREME cooling, Intel tried to push it to 1.13Ghz and then had to recall it as the review samples they'd sent out didn't work properly. :p)

the "Tualatin" Pentium 3, on .13 (with twice the L2 cache, 512K) tops out at ~1.7Ghz (though Intel cut the P3 off at 1.4Ghz to avoid making the P4 look too pathetic ;))

the Tualatin is a more complex design due to the extra cache, and yet it gained ~700Mhz of headroom from the die shrink.

If the GPUL is.. half a Power4, and it's on a smaller process, I'd expect it to massively gain headroom.

As for heat verses clock, it works something like
smaller process = generates less heat at whatever clockspeed BUT then gets clocked higher which balances out the overall heat output.
 
Originally posted by suzerain
To me, "based on" means, pretty specifically, that it is NOT a Power 4. Otherwise they would have just said, "We're announcing a new version of the Power4 for workstations."

It's obviously different enough that they bothered to make that distinction.

Or it could just be that they want to label it as a PowerPC chip, rather than a POWER chip?


Since the POWER4 is a PowerPC-AS chip anyway, could it be that the distinction between the two is merely the intended market?

Ah well, we'll know more on the 15th.

(in other news, did the macrumors server just start a nightly cron job to back everything up or something?.. it's getting very unresponsive from where I'm sitting.)
 
The POWER4 is more powerful than the Itanium 2 in the real world. Furthermore Spec isn't the greatest benchmark of all times and in practice when you are looking at picking between those two chips people rarely base their purchasing decision on those numbers (you'd be pretty dumb if you did).

Keep in mind Itanium is currently an evolution ahead of the POWER4. The Itanium is in the early stages still of its 2nd evolution while the POWER4 is in the very late stages of its 1st. The next evolution of it will clock at around 1.8 GHz, which isn't a long way off the 2GHz of what is coming Oct 15.

All that said what we will see Oct 15 won't be a POWER4. What I expect it will be is a core very similar to the POWER4 core with less L2 cache, a SIMD unit and a new memory subsystem. I think they'd be silly not to plan dual core but they may have other factors preventing them.

It's a different enough chip that you would say it is POWER4 derived and certainly would be incredibly powerful. Most of the cost reduction would come through just removing the large caches the POWER4 uses.
 
The POWER4 is more powerful than the Itanium 2 in the real world. Furthermore Spec isn't the greatest benchmark of all times and in practice when you are looking at picking between those two chips people rarely base their purchasing decision on those numbers (you'd be pretty dumb if you did).

Indeed, but there isn't really anything that qualifies as a "perfect" benchmark, and unless someone has any better suggestions, SPEC serves as a halfway decent "it's fast.. no.. it's slow.. no, wait, it's fast!" test.. even if it is rather compiler dependent :)

All that said what we will see Oct 15 won't be a POWER4. What I expect it will be is a core very similar to the POWER4 core with less L2 cache, a SIMD unit and a new memory subsystem. I think they'd be silly not to plan dual core but they may have other factors preventing them.

The top factor there would be die size/yield

Plus, the interconnects required to keep a pair of cores properly fed with data would be.. excessively complex for a desktop part (high pin count on the processor package, large number of traces on the mainboard.. etc)

It's a different enough chip that you would say it is POWER4 derived and certainly would be incredibly powerful. Most of the cost reduction would come through just removing the large caches the POWER4 uses.

the L3 cache on the POWER4 is external, although 128MB of EDram can't be cheap.

the majority of the price reduction for the processor itself will come from

1) less complex design (single core, less cache on die)
2) process size reduction.

Any guesses at L3 cache speeds for GPUL?, I'm thinking Powermacs with 4MB of DDR Sram at 1/4th the clockspeed? (or perhaps 1/3rd?)
 
Unobtainium vs. G5/Power4...

I don't know that it's really just as simple as all this discussion.

The G4 has a 64-bit core, a 64-bit data bus, a 36-bit (64G?! John Holmes got nuttin' on you!) address bus (along with a bunch of other specialized control doodadz), 128-bit path to L3, a 128-bit butt whoopin math chip - and an arbiter chip between the two CPUs in the current systems fat piped to RAM, where the 2.7GB/s bandwidth is shared with all the fancy new schmangled DMA (A dirty trick SGI taught everybody with their Revolution in Octane arch.) devices hung off an UberBridge.

The G4 on the CPU side is not going any further 'physically' without major chicanery.

All WPs (with the exeption of Book E which we don't care about (what's that for? WAPtv?) on G5 were pulled from HelloMoto well over a year ago (anyone remember the writs on the working protos with the skinny mic fab?) - speculatively, since this is a rumor site, due to issues with getting RapidIO to play well with AltiVec and MPX. Second runner up might be that their having difficulties with running G3/G4 user mode binaries.

I'd bet they do it, and I wouldn't count HelloMoto out just yet. Too much of the ground work is already lain, and the big strength of the Power4 lies in cascading near infinite arrays of CPUs (anyone remember adding mass quantities of 604's in AS/400s?). This doesn't do us much good when we're using AMD's commodity memory architecture.

Apple could definitely do some magic with BSD/Mach, but they need to be able to continue to make cheap computers for cheapo customers. I predict that you may see a stress fissure turn into a rift were OS X gets ported to some high dollar warez - but the consumer side can never get away from widely available commodity sub-parts... same ones Wintel board designers are using.

Apple used to try and compete by making premium warez, but the Great God Cheapoquatl kicked them in the junk.

If you've read this far you're crazy as a loon. Thank you.
c
 
Re: Unobtainium vs. G5/Power4...

Originally posted by zerokelvin
The G4 on the CPU side is not going any further 'physically' without major chicanery.

They aren't trying to take the <b>G4</b> any further, they are looking to replace that router-controller-with-an-oversized-ego completely with something derived from a high-end mainframe chip.

Or is that chicanery? :p

BTW, the POWER4 might scale well in an SMP environment, but it's still damned impressive on a per-processing-core basis, hopefully the GPUL will preserve that aspect of it's performance.
 
Originally posted by Chryx
Indeed, but there isn't really anything that qualifies as a "perfect" benchmark, and unless someone has any better suggestions, SPEC serves as a halfway decent "it's fast.. no.. it's slow.. no, wait, it's fast!" test.. even if it is rather compiler dependent :)
No there are no perfect benchmarks beyond testing the system specifically for task you have planned. Of course when testing desktops you usually have dozens of applications planned and performance can vary.

I have a particular dislike for SPEC though. I've seen too many funny things done with it to really have any faith in it.

The top factor there would be die size/yield

Plus, the interconnects required to keep a pair of cores properly fed with data would be.. excessively complex for a desktop part (high pin count on the processor package, large number of traces on the mainboard.. etc)
Die size wouldn't be too big an issue. It would be around the size of the initial PIV's probably.

I agree the main disadvantage with a dual core chip would be yield. If one core doesn't work the chip is no longer dual core. You can sell it as a single core processor but still you've just wasted money adding the second core that doesn't work. That's really what I expect would be the driving force behind a decision not to produce a dual core GPUL.

Memory handling/interconnects are another problem but in all honesty I wouldn't consider that the worst.

On the other hand from a design and cost point of view there can be substantial benefits in moving to multi-core chips. Over the coming years chip manufactures are increasingly going to be talking about it for desktop chips (Intel I know are looking at it and IBM's views are well known).

If you were going to begin with any chip it would make sense to do it with one derived from a dual core chip as well.

the L3 cache on the POWER4 is external, although 128MB of EDram can't be cheap.

the majority of the price reduction for the processor itself will come from

1) less complex design (single core, less cache on die)
2) process size reduction.

Any guesses at L3 cache speeds for GPUL?, I'm thinking Powermacs with 4MB of DDR Sram at 1/4th the clockspeed? (or perhaps 1/3rd?)
I didn't so much mean with respect to the L3 cache itself, although it is expensive.

The L3 controller and a lot of the circuitry due to the complexity of the POWER4's memory system would be simplified in a desktop chip.

The question of L3 cache is an interesting one as it's the area I would expect is most open to change. I doubt 4MB early on although I expect it will probably handle 2 - 8MB and I'd assume 1/3rd clockspeed.
 
Originally posted by Telomar
No there are no perfect benchmarks beyond testing the system specifically for task you have planned. Of course when testing desktops you usually have dozens of applications planned and performance can vary.

I have a particular dislike for SPEC though. I've seen too many funny things done with it to really have any faith in it.

Intel has a nasty habit of optimising their compilers to specifically perform better in SPEC. One can only assume that such funny business hurts performance in other areas at the expense of better spec marks. All the other big iron processors tend to have spec scores submitted from boxes that use common tools.. like gcc to compile the spec suite. ...at least that is what I've always heard.
originally posted by suzerain
However, I didn't "twist" anything you said; I just quoted it.
Yes, you just reposted what I said, but you posted my comments out of context. Your previous post started with a quote from me stating that 'I guarantee that the GPUL will be faster'. You then went on to argue that the GPUL won't be faster than Itanium. My quote was in direct reference to the post I was commenting on, in which the poster asserted that the upcomming IBM processor won't be faster than the current G4. This is a VERY different argument from the one that you insinuated I was making. In fact, I'm pretty sure that I never made any claims about the upcomming IBM chip in comparison to any processor except the G4 and possibly existing 32bit x86 processors.

I'm just trying to be accurate here. :D

originally posted by zerokelvin

The G4 has a 64-bit core, a 64-bit data bus, a 36-bit (64G?! John Holmes got nuttin' on you!) address bus (along with a bunch of other specialized control doodadz), 128-bit path to L3, a 128-bit butt whoopin math chip - and an arbiter chip between the two CPUs in the current systems fat piped to RAM, where the 2.7GB/s bandwidth is shared with all the fancy new schmangled DMA (A dirty trick SGI taught everybody with their Revolution in Octane arch.) devices hung off an UberBridge.

Owch, hold on. The G4 is considered a 32bit processor. It has 32 bit registers for integers. It can do double precision floats but so can 32bit x86 processors. I certainly wouldn't call the G4 a 64 bit chip.

It also doesn't have a 38 bit address BUS, it has address registers that are 38 bits in size, meaning it could theoretically address quite a bit of memory (though I don't know how to do this in a 32bit ansi C program.

As for the 128 bit butt whompin math chip, it has SIMD registers that are 128 bits wide but the whole point of SIMD is that you perform the same operation on multiple data. You have to have 128 bits of data, at the same time, that needs the same operation done on it. SSE2 on the P4 is also 128 bits wide too.

And, you're being way to nice with your 'fat piped to RAM' comment. The G4 is still only available with a 64bit SDR path to the main chipset and therefore only 64bit SDR access to main system memory.

I'd say that the G4 certainly 'ain't all that' like you seem to think it is. :-(
It's a nice, elegant chip, but it isn't where it should be at this stage in the evolution of the CPU.
 
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.