While CES 2020 technically doesn’t wrap up for another couple of days, I’ve already been asked a good dozen or so times what the neatest or most surprising product at the show has been for me. And this year, it looks like that honor is going to Intel. Much to my own surprise, the company has their first Xe discrete client GPU, DG1, at the show. And while the company is being incredibly coy on any actual details for the hardware, it is none the less up and running, in both laptop and desktop form factors.

Intel first showed off DG1 as part of its hour-long keynote on Monday, when the company all too briefly showed a live demo of it running a game. More surprising still, the behind-the-stage demo unit wasn’t even a desktop PC (as is almost always the case), but rather it was a notebook, with Intel making this choice to underscore both the low power consumption of DG1, as well as demonstrating just how far along the GPU is since the first silicon came back to the labs in October.

To be sure, the company still has a long way to go between here and when DG1 will actually be ready to ship in retail products; along with the inevitable hardware bugs, Intel has a pretty intensive driver and software stack bring-up to undertake. So DG1’s presence at CES is very much a teaser of things to come rather than a formal hardware announcement, for all the pros and the cons that entails.

At any rate, Intel’s CES DG1 activities aren’t just around their keynote laptop demo. The company also has traditional desktop cards up and running, and they gave the press a chance to see it yesterday afternoon.

Dubbed the “DG1 Software Development Vehicle”, the desktop DG1 card is being produced by Intel in small quantities so that they can sample their first Xe discrete GPU to software vendors well ahead of any retail launch. ISVs regularly get cards in advance so that they can begin development and testing against a new GPU architecture, so Intel doing the same here isn’t too unusual. However it’s almost unheard of for these early sample cards to be revealed to the public, which goes to show just how different of a tack Intel is taking in the launch of its first modern dGPU.

Unfortunately there’s not much I can tell you about the DG1 GPU or the card itself. The purpose of this unveil is to show that DG1 is here and it works; and little else. So Intel isn’t disclosing anything about the architecture, the chip, power requirements, performance, launch date, etc. All I know for sure on the GPU side of matters is that DG1 is based on what Intel is calling their Xe-LP (low power) microarchitecture, which is the same microarchitecture that is being used for Tiger Lake’s Xe-based iGPU. This is distinct from Xe-HP and Xe-HPC, which according to Intel are separate (but apparently similar) microarchitectures that are more optimized for workstations, servers, and the like.

As for the dev card, I can tell you that it is entirely PCIe slot powered – so it’s under 75W – and that it uses an open-air cooler design with a modest heatsink and an axial fan slightly off-center. The card also features 4 DisplayPort outputs, though as this is a development card, this doesn’t mean anything in particular for retail cards. Otherwise, Intel has gone with a rather flashy design for a dev card – since Intel is getting in front of any dev leaks by announcing it to the public now, I suppose there’s no reason to try to conceal the card with an unflattering design, which instead Intel can use it as another marketing opportunity.

Intel has been shipping the DG1 dev cards to ISVs as part of a larger Coffee Lake-S based kit. The kit is unremarkable – or at least, Intel wouldn’t remark on it – but like the DG1 dev card, it’s definitely designed to show off the video card inside, with Intel using a case that mounts it parallel to the motherboard so that the card is readily visible.

It was also this system that Intel used to show off that, yes, the DG1 dev card works as well. In this case it was running Warframe at 1080p, though Intel was not disclosing FPS numbers or what settings were used. Going by the naked eye alone, it’s clear that the card was struggling at times to maintain a smooth framerate, but as this is early hardware on early software, it’s clearly not in a state where Intel is even focusing on performance.

Overall, Intel has been embarking on an extended, trickle-feed media campaign for the Xe GPU family, and this week’s DG1 public showcase is the latest in that Odyssey campaign. So expect to see Intel continuing these efforts over the coming months, as Intel gradually prepares Tiger Lake and DG1 for their respective launches.

Comments Locked

83 Comments

View All Comments

  • zmatt - Thursday, January 9, 2020 - link

    Hell, you can even read Anand's review from 1998 still. https://www.anandtech.com/show/204
  • Tomatotech - Thursday, January 9, 2020 - link

    Oh wow that was a wild read. I was going to say they tested it with Doom, but no, they tested it on Win-95 with Quake 2 at 640x480, getting a whopping 21 frames per second.

    Top performer was the Voodoo 2 GPU, with a whole 12MB of on-card RAM, getting 48fps at 800x600. Comparing that with the struggles a 2080 ti has with ray-tracing at 4k, some framerates never change...
  • Foeketijn - Friday, January 10, 2020 - link

    Thank you for finding that article. That's from before I even knew Anandtech. The CPU world got a bit more interesting lately. Maybe the GPU's will be revived also.
  • Dolda2000 - Thursday, January 9, 2020 - link

    Frankly, I'm just glad they're actually showing off things still in development instead of keeping everything under complete wraps until it's completely ready for final release. That should be encouraged, not met with "oh wow, look how much this half-finished project sucks right now".
  • DasGanon - Thursday, January 9, 2020 - link

    It's not 7 years, closed beta was 7 years ago. It's GaaS that has had a lot of changes since then (and depending on game mode even an I7-4790K & GTX 970 has some hiccups, never below 60fps but it's not consistent)

    Is it a great demo, no because Warframe is one of the weirdly best optimized games ever and it was extremely recently WF got rid of XP support.
  • BigDragon - Thursday, January 9, 2020 - link

    You're right -- Warframe has had a lot of changes. DE recently made a big change to the lighting engine that could impact performance. The game is wildly optimized. Integrated graphics on potato PCs can run Warframe buttery smooth. Instead, Intel is trying to build hype with https://youtu.be/gXVRSCQ9zGo . I'm unwilling to accept Intel's excuse that immature hardware and drivers cannot handle Warframe.
  • Cullinaire - Thursday, January 9, 2020 - link

    It's optimized... for AMD and nvidia.
  • Spunjji - Friday, January 10, 2020 - link

    That's not how any of this works.

    It's possible to tailor a game for one specific architecture by implementing features that play to that architecture's strengths - e.g. Hairworks, TressFX, overdoing tesselation, etc. That will make it run well for that vendor or product at the expense of everyone else's. The rest of the optimization done for specific hardware is done at the driver level by the vendor, though.

    There's not really any way that Warframe itself could be "optimized" for both AMD and Nvidia but not Intel. It's their job to make existing 6-year-old games run well on it, not developers.
  • Spunjji - Friday, January 10, 2020 - link

    That's what I thought, too. Warframe runs at 60fps in 1080p with medium settings on lightly optimized 2400G systems. I get that this is only an early demo... but it's a terrible one.
  • AshlayW - Sunday, January 12, 2020 - link

    To be fair it has received some major graphics overhaul. But yeah. It's not demanding. My 3400g does just fine

Log in

Don't have an account? Sign up now