Sony just announced the PlayStation 4, along with some high level system specifications. The high level specs are what we've heard for quite some time:

  • 8-core x86-64 CPU using AMD Jaguar cores (built by AMD)
  • High-end PC GPU (also built by AMD), delivering 1.84TFLOPS of performance
  • Unified 8GB of GDDR5 memory for use by both the CPU and GPU with 176GB/s of memory bandwidth
  • Large local hard drive

Details of the CPU aren't known at this point (8-cores could imply a Piledriver derived architecture, or 8 smaller Jaguar cores—the latter being more likely), but either way this will be a big step forward over the PowerPC based general purpose cores on Cell from the previous generation. I wouldn't be too put off by the lack of Intel silicon here, it's still a lot faster than what we had before and at this level price matters more than peak performance. The Intel performance advantage would have to be much larger to dramatically impact console performance. If we're talking about Jaguar cores, then there's a bigger concern long term from a single threaded performance standpoint.

Update: I've confirmed that there are 8 Jaguar based AMD CPU cores inside the PS4's APU. The CPU + GPU are on a single die. Jaguar will still likely have better performance than the PS3/Xbox 360's PowerPC cores, and it should be faster than anything ARM based out today, but there's not huge headroom going forward. While I'm happier with Sony's (and MS') CPU selection this time around, I always hoped someone would take CPU performance in a console a bit more seriously. Given the choice between spending transistors on the CPU vs. GPU, I understand that the GPU wins every time in a console—I'm just always an advocate for wanting more of both. I realize I never wrote up a piece on AMD's Jaguar architecture, so I'll likely be doing that in the not too distant future. Update: I did.

The choice of 8 cores is somewhat unique. Jaguar's default compute unit is a quad-core machine with a large shared L2 cache, it's likely that AMD placed two of these together for the PlayStation 4. The last generation of consoles saw a march towards heavily threaded machines, so it's no surprise that AMD/Sony want to continue the trend here. Clock speed is unknown, but Jaguar was good for a mild increase over its predecessor Bobcat. Given the large monolithic die, AMD and Sony may not have wanted to push frequency as high as possible in order to keep yields up and power down. While I still expect CPU performance to move forward in this generation of consoles, I was reminded of the fact that the PowerPC cores in the previous generation ran at very high frequencies. The IPC gains afforded by Jaguar have to be significant in order to make up for what will likely be a lower clock speed.

We don't know specifics of the GPU, but with it approaching 2 TFLOPS we're looking at a level of performance somewhere between a Radeon HD 7850 and 7870. Update: Sony has confirmed the actual performance of the PlayStation 4's GPU as 1.84 TFLOPS. Sony claims the GPU features 18 compute units, which if this is GCN based we'd be looking at 1152 SPs and 72 texture units. It's unclear how custom the GPU is however, so we'll have to wait for additional information to really know for sure. The highest end PC GPUs are already faster than this, but the PS4's GPU is a lot faster than the PS3's RSX which was derived from NVIDIA's G70 architecture (used in the GeForce 7800 GTX, for example). I'm quite pleased with the promised level of GPU performance with the PS4. There are obvious power and cost constraints that would keep AMD/Sony from going even higher here, but this should be a good leap forward from current gen consoles.

Outfitting the PS4 with 8GB of RAM will be great for developers, and using high-speed GDDR5 will help ensure the GPU isn't bandwidth starved. Sony promised around 176GB/s of memory bandwidth for the PS4. The lack of solid state storage isn't surprising. Hard drives still offer a dramatic advantage in cost per GB vs. an SSD. Now if it's user replaceable with an SSD that would be a nice compromise.

Leveraging Gaikai's cloud gaming technology, the PS4 will be able to act as a game server and stream the video output to a PS Vita, wirelessly. This sounds a lot like what NVIDIA is doing with Project Shield and your NVIDIA powered gaming PC. Sony referenced dedicated video encode/decode hardware that allows you to instantaneously record and share screenshots/video of gameplay. I suspect this same hardware is used in streaming your game to a PS Vita.

Backwards compatibility with PS3 games isn't guaranteed and instead will leverage cloud gaming to stream older content to the box. There's some sort of a dedicated background processor that handles uploads and downloads, and even handles updates in the background while the system is off. The PS4 also supports instant suspend/resume.

The new box heavily leverages PC hardware, which is something we're expecting from the next Xbox as well. It's interesting that this is effectively how Microsoft entered the console space back in 2001 with the original Xbox, and now both Sony and MS have returned to that philosophy with their next gen consoles in 2013. The PlayStation 4 will be available this holiday season.

I'm trying to get more details on the CPU and GPU architectures and will update as soon as I have more info.

Source: Ustream

Comments Locked

160 Comments

View All Comments

  • gruffi - Thursday, February 21, 2013 - link

    C'mon, get your facts right. Jaguar is a low-power core. It has excellent performance/watt and performance/mm^2. There is nothing from Intel which can compete with it. Atom is too slow, SB/IB is too big and too expensive. We are NOT talking about the big server design from AMD, Orochi. And even the Bulldozer cores are not "bigger/noisier" than comparable Intel cores.
  • plonk420 - Thursday, February 21, 2013 - link

    amd (bulldozer) vs intel performance (sandy):

    "(pretty much) the same"
    AvP
    BF3
    Battleforge
    Crysis 2
    DA2
    Hard Reset (1600p)
    Metro 2033
    Stalker: COP (1600p)
    Shogun 2 (1600p)
    WoW Pre-Pandaria? (1600p)
    3DMark11
    Unigine Heaven 2.0

    "similar"
    Arkham City
    Dirt 3
    Stalker: COP (1200p)
    Shogun 2 (1200p)
    WoW Pre-Pandaria? (1200p)

    "pitifully slower"
    Civ5
    Hard Reset (1200p)
    SC2
    Skyrim
  • plonk420 - Thursday, February 21, 2013 - link

    oh, sauce is techPowerUp
  • xaml - Friday, February 22, 2013 - link

    Sauce?
  • sohcermind - Sunday, February 24, 2013 - link

    sauce = source that's how you say it in a New York accent.
  • Exodite - Thursday, February 21, 2013 - link

    To be fair the Bulldozer vs. Sandy Bridge comparison is rather irrelevant given that this is about Jaguar.

    Anyway, Intel's single-threaded performance is significantly higher but that's only reflected in games that have computationally intensive threads.

    Usually MMOs, RPGs and tactical games.

    Shooters and other action-genre games traditionally aren't as CPU-limited, which means it shouldn't impact console gaming as much as it would PC gaming.

    That said, the per-thread performance of the PS4's Jaguar cores are likely to be significantly slower than that of the Bulldozer chip used for the comparison.
  • Krysto - Thursday, February 21, 2013 - link

    It's not even one of the high-end CPU's though. It's Jaguar cores, which are the successor of Bobcat (competitor to Atom). Jaguar seems about as powerful as ARM's Cortex A15 to me. Not sure why they'd prefer 8 of those over 4 high-end cores. Just because of price?

    Also I hope the GPU is at least a GCN one. Can we confirm that?
  • Oldboy1948 - Saturday, February 23, 2013 - link

    In Geekbench no ARMs have any results in vector perf. Why?

    For Bobcat it is very much about size and speed of cache. E-350 has L2 running at half speed. This is much too slow. Full speed is a must. 8 cores at 2GHz with full speed 4MB L2.
    The L1 cache for Bobcat is much better than Bulldozer's. Separate L1 for every core must be an optimum. Bulldozer should have 64i+64d for every core...
  • cjb110 - Thursday, February 21, 2013 - link

    Remember that the Devs will be a lot closer to the 'wire', and wont have Windows in the way, so AMD's disadvantage in Windows might not be as significant. Carmack has repeatedly pushed for MS to allow game devs closer access to the hardware.

    Plus cost, if your loosing 20% performance AND 50% cost...then it makes sense for a mass consumer device.
  • ionis - Thursday, February 21, 2013 - link

    I wonder what the OS will be. Given Sony's history, I doubt it'll be Linux/BSD. It's probably not going to be Windows or Mac. So what does that leave? Did they make their own x86 OS?

Log in

Don't have an account? Sign up now