Xbox One (Durango) Technical hardware investigation

Status
Not open for further replies.
I think the news appeared in an article first. According to it, Microsoft took the unprecedented decision to upgrade the specs of the Xbox One before its release.

This is the actual article from July 12th or so, and it mentions the survey sent out to developers, a rumoured upclock -which turned out to be real-, how they took the decision due to the DRM issues and an alleged upgrade to 12GB because development kits shipped with 12GBs then --although developers thought retail kits would have 8GB.

http://www.extremetech.com/gaming/1...-the-specs-of-the-xbox-one-before-its-release

Its been floating around these forums since atleat Febuary.. Guilty as charged im afraid.. http://forum.beyond3d.com/showpost.php?p=1714023&postcount=2955
 
I think the best bet for a ram upgrade would be to replace the ram in the xbox with higher density chips/sticks and hope the firmware supports the upgrade.

Because odds are it's the OS reserve that is specified and not the ram the game uses.

I'm sure this console season will have unprecedented box modding due to the standardized nature of the hardware.
 
I think the news appeared in an article first. According to it, Microsoft took the unprecedented decision to upgrade the specs of the Xbox One before its release.

This is the actual article from July 12th or so, and it mentions the survey sent out to developers, a rumoured upclock -which turned out to be real-, how they took the decision due to the DRM issues and an alleged upgrade to 12GB because development kits shipped with 12GBs then --although developers thought retail kits would have 8GB.

http://www.extremetech.com/gaming/1...-the-specs-of-the-xbox-one-before-its-release

Did you actually read the articles? Because they completely contradict each other and any argument of 12 GB.

Extremetech said:
The RAM rumor seems to have originated from the Xbox One dev kits shipping with 12GB of RAM. However, the source claims that the retail kits do not pack that much RAM.

Examiner said:
Physical RAM won't be upped in time for November release as it was too late even during initial reveal...When asked about the increase in RAM (from 8GB to 12GB), the response I obtained was simply "I have heard nothing like that, the development kits have 12GB of RAM, but as far as I know the retail kits are locked in at 8GB."

Link to Extremetech article

Link to Examiner article
 
Did you actually read the articles? Because they completely contradict each other and any argument of 12 GB.





Link to Extremetech article

Link to Examiner article

Both articles say that Dev Kits have 12GB of RAM. And, if true, that mean 12GB is (technically) feasible, not impossible at all .

EDIT: This is interesting (from the examiner article):

As for current projects, they also are "too far gone for a bump to be beneficial" as many developers have been developing for the currently announced set of Xbox One specifications. He did mention that the possibility may still be there though, as the increase in RAM would help development of games 2014 and beyond when taking second screen gaming into consideration - something many developers have been focusing on.
 
Last edited by a moderator:
The original dev kits (PCs) had 12GB which was most likely a triple channel configuration like 4GBx3DIMMs. There is no info on dev kits based on the final XB1 hardware.
 
Apparently the speed upgrade to the GPU was not a hardware one but a change to the OS similar to the WiiU I guess. Does this mean that it is just an overclock not an actual upgrade? And if you can do that through the OS then how long before someone hacks that to push the envelope?

Source
 
Apparently the speed upgrade to the GPU was not a hardware one but a change to the OS similar to the WiiU I guess. Does this mean that it is just an overclock not an actual upgrade? And if you can do that through the OS then how long before someone hacks that to push the envelope?

Source

On most of these devices, clocks are based on some multiple of some base clock, selected in the firmware, or bios.
Depending on the device various components will have different clock domains. So the GPU clock may or may not be tied to the ESRAM clock or the CPU clock.
 
Apparently the speed upgrade to the GPU was not a hardware one but a change to the OS similar to the WiiU I guess. Does this mean that it is just an overclock not an actual upgrade? And if you can do that through the OS then how long before someone hacks that to push the envelope?

Source

Someone correct me if I'm wrong but im pretty sure that clocks for at least any modern system are tied to some software (BIOS/firmware, etc) and not to physical hardware. I can imagine that dynamic clocking could actually be useful for them to a degree to keep sound and heat down in situations were games are not open.
 
Someone correct me if I'm wrong but im pretty sure that clocks for at least any modern system are tied to some software (BIOS/firmware, etc) and not to physical hardware. I can imagine that dynamic clocking could actually be useful for them to a degree to keep sound and heat down in situations were games are not open.

Yes of course. It's the same as how you'd overclock your CPU on your PC. You go into the bios.

I think that's what the MS person meant, I guess they are somewhat correct but it's more like a firmware change. They are just simplifying for a mass audience and/or dont fully understand it themselves.

Of course a chip's clockspeed is not some sort of physical property of it as it rolls off the line...
 
Clocks can entirely be controlled by the OS layer if need be, the only elements that are set at the hardware level (which will be controlled by the PCB) is a default "boot" voltages and speeds which may not necessarily relate to any clocks that you would think of as running clocks.

Our discrete GPU's actually have a microcontroller on them with arbitration code that dictates what clock/voltage state it should be operating at dependant on a number of parameters - i.e. if the activity counters on the chip are high it will push to a high clock state for peak performance, or if the activity is low it will drop to a lower clock/voltage to save power; alternatively if it is calculating that the activity is hitting a power threshold then it will modulate between states to keep the power in check. The microcode to do this is loaded into the GPU from a table in the BIOS when the driver is loaded by the OS but those limits can further be controlled by software in the OS.

So when they are saying "the GPU is clocked higher" but also saying "we've not changed any hardware" these don't contradict each other; likelihood is that whatever part of the system dictated the max clock / voltage state of the GPU portion (be that firmware or OS) has just been updated.
 
Yeah there are utilities to overclock CPU/GPU from Windows nowdays on PC.

I guess it doesn't matter, point is more like "parameters get loaded on boot" whether "bios" or OS.
 
Clocks can entirely be controlled by the OS layer if need be, the only elements that are set at the hardware level (which will be controlled by the PCB) is a default "boot" voltages and speeds which may not necessarily relate to any clocks that you would think of as running clocks.

Our discrete GPU's actually have a microcontroller on them with arbitration code that dictates what clock/voltage state it should be operating at dependant on a number of parameters - i.e. if the activity counters on the chip are high it will push to a high clock state for peak performance, or if the activity is low it will drop to a lower clock/voltage to save power; alternatively if it is calculating that the activity is hitting a power threshold then it will modulate between states to keep the power in check. The microcode to do this is loaded into the GPU from a table in the BIOS when the driver is loaded by the OS but those limits can further be controlled by software in the OS.

So when they are saying "the GPU is clocked higher" but also saying "we've not changed any hardware" these don't contradict each other; likelihood is that whatever part of the system dictated the max clock / voltage state of the GPU portion (be that firmware or OS) has just been updated.

Out of interest and totally unrelated to any console : is it possible to disable/enable CUs via software?
 
Out of interest and totally unrelated to any console : is it possible to disable/enable CUs via software?

I'm not Dave but, you used to be able to buy a few GPU's and get more execution units out of them (they weren't CU's back then) by flashing them with certain Bioses IIRC. X800 was one back in the day...which would suggest it's possible.

Now I think they "laser cut" off any units they dont want you to use, so they're basically hardware blocked, on PC. But that's to prevent people gaming the system, not sure if they'd do that on a console.
 
Apparently the speed upgrade to the GPU was not a hardware one but a change to the OS similar to the WiiU I guess. Does this mean that it is just an overclock not an actual upgrade? And if you can do that through the OS then how long before someone hacks that to push the envelope?

Source

You mean developers actually hacking the X1, pushing clock speeds without Microsoft's permission and possibly losing their license to develop for the platform? You really think they'd take a risk like that?

If anything that article represents yet another example of the kinds of sites that are covering gaming and gaming technology. This Adam Barnes doesn't seem to have a single clue of how upclocking on a modern system occurs. Hmmmm.
 
I'm not Dave but, you used to be able to buy a few GPU's and get more execution units out of them (they weren't CU's back then) by flashing them with certain Bioses IIRC. X800 was one back in the day...which would suggest it's possible.

Now I think they "laser cut" off any units they dont want you to use, so they're basically hardware blocked, on PC. But that's to prevent people gaming the system, not sure if they'd do that on a console.

Well didn't Sony disable SPEs on the CELL?
 
Looks like AMD will unveil Hawaii/Volcanic Islands tech in late September, launch at Q4.

http://semiaccurate.com/2013/08/07/amd-to-launch-hawaii-in-hawaii/

A little quote:

This is a massively updated GPU vs the minor tweaks in the last round. Given the timing, Hawaii is unquestionably a 28nm part so no shrink related performance bump but the architectural changes should more than make up for that. It is unlikely to be an incremental advance.


I'm posting this here because I'm wondering if it's possible Durango (or Orbis) are fully/partially based on new tech, not Bonaire nor Pitcairn, just like Xenon had Unified Shaders more than 6 month earlier the PC product was launched. At least better HSA integration or efficiency improvements?

At Q4, AMD 7970 will be 2 years old.
 
You mean developers actually hacking the X1, pushing clock speeds without Microsoft's permission and possibly losing their license to develop for the platform? You really think they'd take a risk like that?

If anything that article represents yet another example of the kinds of sites that are covering gaming and gaming technology. This Adam Barnes doesn't seem to have a single clue of how upclocking on a modern system occurs. Hmmmm.

Not developers. Regular joe hacks who like to tinker with hardware or is it more likely that MS will lock that feature down in production?

I get the way in which CPU\GPU's are upclocked but there were a few people who were talking about hardware respins etc And this kind of knocks that on the head.
 
I get the way in which CPU\GPU's are upclocked but there were a few people who were talking about hardware respins etc And this kind of knocks that on the head.
To be fair it doesn't actually knock those on the head.
The respins may have been what got the gpu to 800-853.

The setting it in 'software', just means that they could settle on the final clock speed once all the validation, TDP, yield results, etc had been taken into account.
 
Status
Not open for further replies.
Back
Top