That's why I am telling you that this information and more are disclosed in different avenues that are under NDA. You are not critically thinking if you think that Sony would be communicating key information about hardware to developers via public presentations like the one Cerny did.
There are other NDA communication channels for the lower end developers. That is why this claim of it being for developers is not true. All authorized PlayStation developers would have access to this information prior to Cerny's presentation with significantly more detail but all under NDA.
Oh really?
PS4 Developers Didn't Know About the Console's 8GB GDDR5 RAM.
While February’s big PlayStation 4 coming out party prompted plenty of gasps around the globe, the sharpest intake of breath came when system architect Mark Cerny confirmed that the next generation system would ship with 8GB GDDR5 RAM. However, the revelation didn't just shake the media and fans alike – it also came as a complete surprise to large pockets of the development community, who had no idea how much memory the manufacturer was packing into the impending console.
"We were told [the PS4] was 4GB originally," Just Add Water’s CEO Stewart Gilray told VideoGamer.com. “We first knew it had 8GBs when Mark said at the event's stage, 'And it has 8GB of memory.' We'd had kits at that point for a good while."
@tormentos: You're a lost cause, I await for 5yrs down the road where you cut out my whole comment to a snippet that makes me look bad.
Enjoy those 9.2TF in 2025 when you can afford one LMFAO.
@Livecommander said:
@i_p_daily: I dont remember you making fun of the xbone for having less power. And it was more expensive too.
The new gen differences are going to be way smaller and 3rd parties will preform better on ps5 unless you think an occasional 5fps jump is more impactful than instant asset loading.
And we all know that Sony is the graphics king when It comes to exclusives.
I don't care about power, never have, never will.
Its you cows who care, and you cows lose your shit when Sony loses ANYTHING to MS, just like you and tormy are, that's why I will make fun of you losers lol.
1. ROPS debates are useless without memory bandwidth debate.
2. Memory bandwidth debate is pointless without CU power debate. CU itself has TMU I/O.
There are 8 RB units for 64 ROPS.
10 RB units x 8 ROPS yields 80 ROPS. One extra RB unit for each Shader Engine cluster.
Techpowerup speculated that each RB unit has a GDDR6 chip.
320 bit / 256 bit = 1.25
1.25 x 64 ROPS = 80 ROPS.
The xbox one has 16ROP and you could not stop damage controlling it.
Hell you still do..lol
@i_p_daily said:
@tormentos: You're a lost cause, I await for 5yrs down the road where you cut out my whole comment to a snippet that makes me look bad.
Enjoy those 9.2TF in 2025 when you can afford one LMFAO.
@Livecommander said:
@i_p_daily: I dont remember you making fun of the xbone for having less power. And it was more expensive too.
The new gen differences are going to be way smaller and 3rd parties will preform better on ps5 unless you think an occasional 5fps jump is more impactful than instant asset loading.
And we all know that Sony is the graphics king when It comes to exclusives.
I don't care about power, never have, never will.
Its you cows who care, and you cows lose your shit when Sony loses ANYTHING to MS, just like you and tormy are, that's why I will make fun of you losers lol.
Yes you do which is why you changed accounts into this alt.
And it shows on the first bold sentense there were you insist in using the power advantage of the xbox as some kind of ammo.
Worse in your case because you are inventing shit claiming the PS5 is 9.2TF when is not.
Not that it would change much if it was 9.2TF MS would still get beat up..lol
My argument is not flawed and was backed by several PS4 games that had way a bigger gap than they should relative to the power advantage.
Never the less you downplayed it then just like you are downplaying it now thanks for proving me right.
So double the ROP only matter when the advatage is on xbox.
NAVI 10's RBs (color and z-buffer ROPS) have high speed 4MB L2 cache links which are less bound by external bandwidth when compared to pre-Vega /pre-X1X ROPS.
Vega 64 ROPS gains 4MB L2 cache link. 4MB L2 cache is shared with TMUs.
X1X 32 ROPS gains 2MB render cache link. 2MB L2 cache is assigned to TMUs.
NVIDIA Maxwell architecture introduces ROPS with L2 cache link. GTX 980 Ti's 96 ROPS has ~3MB L2 cache with ~6 TFLOPS. 3MB L2 cache is shared with TMUs.
Most ROPS debates refer to color ROP count. Every color ROP unit has 4 z-buffer ROPS units e.g. 7970 has 32 color ROPS with 128 z-buffer ROPS
NAVI gains "DCC everywhere".
My argument changes when there are major design changes.
Note why PowerColor RX-5600 XT Red Devil OC's 192-bit bus penalty hit was just 8 percent from RX 5700.
PowerColor RX-5600 XT OC and RX 5700 have similar TFLOPS power with different memory bandwidth i.e. 336 GB/s vs 448 GB/s.
AMD is slowly following NVIDIA's improvement path.
For Sony's PS5, 448 GB/s shared memory bandwidth has RDNA's DCC everywhere and ROPS being connected to at least 4MB L2 cache and CPU with at least 8MB L3 cache (from PC Ryzen 4800HS series APU).
They were caught off guard and shocked by MS 12 Teraflops of power. PS5 was only 9.2. So then got desperate and try to overclock the Gpu as much as they could before it broke. That's what Cerny actually said.
They were caught off guard and shocked by MS 12 Teraflops of power. PS5 was only 9.2. So then got desperate and try to overclock the Gpu as much as they could before it broke. That's what Cerny actually said.
From Mark Cerny
"When that worst case game arrives, it will run at a lower clock speed. But not too much lower, to reduce power by 10 per cent it only takes a couple of percent reduction in frequency, so I'd expect any downclocking to be pretty minor," he explains. "All things considered, the change to a variable frequency approach will show significant gains for PlayStation gamers."
1% less clock speed yields 10.173 TFLOPS, hence 19.4% difference from XSX.
2% less clock speed yields 10,070 TFLOPS, hence 20.6% difference from XSX.
...
There could be more clock speed reduction.
Mark Cerny warns CPU's AVX workload can cause clock speed throttle.
Log in to comment